Here's an update from the Chair of the JCP and director of the JCP Program office: Patrick Curran

http://java.ulitzer.com/node/965152


JSR Watch: Here’s to Progress

And here’s to the next 10 years!
By Patrick Curran

May 15, 2009 03:00 PM EDT
Reads: 560

The end of the year is an opportunity to review the past year's activity, and to present this to our Executive Committee (EC) members, to our broader membership, and to the general public. So this month I will summarize our progress during the past year.

PMO Initiatives
First, in addition to the ongoing work of moving JSRs through the process (more on this later), the JCP engaged in a couple of new initiatives around transparency and agility.

I've addressed the transparency issue relatively recently in this column, so I won't say much more here except to remind you that we are now strongly encouraging all Expert Groups (EGs) to work in an open and transparent manner by adopting practices such as the use of public mailing lists and issue-tracking mechanisms. Of course, it would be hypocritical for us to encourage this behavior in EGs while continuing to hold Executive Committee (EC) meetings in private, so there too we are becoming more open. Starting in September 2008 the ECs agreed to make full minutes and meeting materials accessible to the general public rather than simply posting summaries that only JCP members could read. (We reserve the right to go into Private Session from time to time when sensitive matters are discussed, but we don't expect to do this very often.) If you want to see what we're up to, the meeting materials are accessible.

As for agility, when I reviewed 2007 activity this time last year it became apparent that the amount of time it takes Expert Groups to complete their work varies significantly. Some manage to finish in a little more than a year, while others take several years. Also, we know that there are some JSRs that are effectively stalled and really ought to be withdrawn. As a first step to encourage agility we decided to introduce a new category for JSRs that have made no progress for 18 months - these will be labeled as "Inactive" on jcp.org. The PMO will work with the Spec Leads of these JSRs to encourage them to pick up the pace. If it becomes clear that the JSR is unlikely to complete, we will encourage them to withdraw it. In addition, we plan to review all JSRs that reach completion, and others as appropriate, to identify and publicize the good (and bad) practices that affect the speed with which JSRs move through the process.


There's more! Read the rest in J2SE Town...

Views: 29

Comment

You need to be a member of Codetown to add comments!

Join Codetown

Happy 10th year, JCertif!

Notes

Welcome to Codetown!

Codetown is a social network. It's got blogs, forums, groups, personal pages and more! You might think of Codetown as a funky camper van with lots of compartments for your stuff and a great multimedia system, too! Best of all, Codetown has room for all of your friends.

When you create a profile for yourself you get a personal page automatically. That's where you can be creative and do your own thing. People who want to get to know you will click on your name or picture and…
Continue

Created by Michael Levin Dec 18, 2008 at 6:56pm. Last updated by Michael Levin May 4, 2018.

Looking for Jobs or Staff?

Check out the Codetown Jobs group.

 

Enjoy the site? Support Codetown with your donation.



InfoQ Reading List

Cactus v1: Cross-Platform LLM Inference on Mobile with Zero Latency and Full Privacy

Cactus, a Y Combinator-backed startup, enables local AI inference to mobile phones, wearables, and other low-power devices through cross-platform, energy-efficient kernels and a native runtime. It delivers sub-50ms time-to-first-token for on-device inference, eliminates network latency, and defaults to complete privacy.

By Sergio De Simone

Presentation: Ecologies and Economics of Language AI in Practice

Jade Abbott discusses the shift from massive, resource-heavy models to "Little LMs" that prioritize efficiency and cultural sustainability. She explains how techniques like LoRA, quantization, and GRPO allow for high performance with less compute. By sharing the "Ubuntu Punk" philosophy, she shares how to move beyond extractive data practices toward human-centric, sustainable AI systems.

By Jade Abbott

Python Workers Redux: Wasm Snapshots and Native uv Tooling

Cloudflare's latest advancements in Python Workers revolutionize serverless performance with near-instant cold starts, expanded package compatibility, and streamlined workflows via the uv package manager. By leveraging memory snapshots and WebAssembly, Cloudflare drastically reduces startup times, making Python a prime choice for AI and data science applications.

By Steef-Jan Wiggers

Nuxt Introduces Native Request Cancellation and Async Handler Extraction for Performance Gains

Nuxt 4.2 elevates the developer experience with native abort control for data fetching, improved error handling, and experimental TypeScript support. With a 39% reduction in bundle sizes and a streamlined app directory, this release enhances performance and project organization, positioning Nuxt as a leading choice for full-stack web applications built on Vue.js.

By Daniel Curtis

OpenAI and Anthropic Donate AGENTS.md and Model Context Protocol to New Agentic AI Foundation

OpenAI and Anthropic have donated their AGENTS.md and Model Context Protocol projects to the Agentic AI Foundation (AAIF), a new directed fund under the Linux Foundation. Block contributed their agent framework, goose, as another founding project, and several other tech companies have joined as Platinum members.

By Anthony Alford

© 2025   Created by Michael Levin.   Powered by

Badges  |  Report an Issue  |  Terms of Service