I Use This!
Activity Not Available

News

Analyzed 4 months ago. based on code collected 4 months ago.
Posted about 12 years ago by Doug Arthur
The logo contest was a major milestone for our project. The community and PPMC members came together and voted on a logo, and were in-sync for the entire ride. The community and PPMC members both chose Adrian's (#42), via Tomasz Maciag with Fuse ... [More] Collective, and Julien's (#49) design as the top choices in the first round. This was a clear sign that the community and PPMC are in-sync. This contest has shown how thriving and active this community is, and suggests it will continue to be. Not only has this contest been evidence of the cohesion of the community as a unit, but the sheer amount of topics and active community members engaging in these discussions have made this a great project to be a part of. There a lot of awesome discussions going on, and some of those discussions are even a little bit difficult, but they are still incredibly productive. You can view the discussions on the flex-dev mailing list archives. While Adrian's design came out on top in the final round, Julien's took the number one spot with the community for the first round. A very good accomplishment, I must say. Congratulations! The total number of submissions we received was astonishing. The contest received a total of 54 submissions. While some designers submitted multiple logo's, most of the submissions were from unique individuals. In the first round of voting we, received a total of 129 unique votes, with 22 of those from PPMC members. The voting process was to distribute up to 5 points between at least 2 logos. In the first round, Julien's design received a total of 106.5 points, and Adrian's design received a total of 99.5 points, with third place taken by Alberto Garcia Ruiz (#40) with 65 points. The PPMC members gave 31 points to Adrian's, and 23 points to Julien's. The complete stats for the first round can be found here. In the second round of voting, we received a total of 126 unique votes, with 21 of those from PPMC members. In this round, it was a simple +1 vote for the logo of choice. Adrian's design came out on top in the second round with a total of 77 votes, while Julien's design had 49 votes. The PPMC members also favored Adrian's design with 16 votes, while Julien's design received 5 votes from PPMC members. The complete stats for the second, and final, round of voting can be found here. You can also find this on the contest page here. Now that the contest has wrapped up, the next step is to potentially make some small tweaks to the logo. Some discussion has been made regarding the color of the logo, and even the placement of the word "apache" in the logo. If you subscribe to the flex-dev mailing list, you can follow along as these developments unfold. These are some exciting times for Apache Flex, and I look forward to being a part of this project. In addition, everyone involved should be proud and excited to be part of this project and its future. [Less]
Posted about 12 years ago by dblevins
It's time! Apache TomEE 1.0.0-beta-2 and Apache OpenEJB 4.0.0-beta-2 are finally out! These releases are major improvements over their predecessors and incorporate heavy doses user feedback and road testing. Nicely balanced at 50% bug fixes, 30% ... [More] improvement of existing features, and 20% filling out features for better production and test experience, Apache TomEE is a major leap toward. Based on feedback of 1.0.0-beta-2, a final release could be right around the corner. Top among the areas to show the most improvement include CDI and JAX-RS, both of which received major attention. Support for an Arquillian adapter for TomEE Embedded is a must-not-miss new feature. The project itself has been taking a test-first approach to support and has amassed a rapidly growing number of tests for all the issues reported during the 1.0.0-beta-1 cycle. An Arquillian adapter for remote Apache TomEE usage does exist as a snapshot and should make it into the next release. As well this release boasts over 70 example applications, making a remarkable resource that cannot be ignored. See the downloads page for the full changelog. Get the binaries while they're hot! Thank you so much to all the users who gave feedback and contributed towards the quality of the release. We're very lucky to have gotten exactly the kind of user feedback were after with the 1.0.0-beta-1 release and we couldn't be prouder of the results. [Less]
Posted about 12 years ago by Sally
The Apache Software Foundation (ASF) Conference Committee has opened a Request for Proposals (RFP) to produce, manage, and execute ApacheCon North America, the official conference of the ASF. Submissions will be accepted through 8 March, 2012. For program requirements and to submit a proposal, please visit http://s.apache.org/oh # # #
Posted about 12 years ago by alg
Apache OpenOffice 3.4 supports embedding SVG graphics using a newly created native SVG interpreter implementation. I want to talk about the advantages and some internals of this solution and the necessary changes done. One reason to do this was ... [More] IP clearance. It allowed removal of six GPL/LGPL libraries, namely librsvg, libcroco, libgsf, gdk-pixbuf, glib, and pango gettext. These were used as an external pixel-based renderer. The new SVG uses an own internal interpreter in a new library and some new UNO API services. IP clearance was no interesting task to do, but it leaded to effects like here with SVG; the install sets get smaller (less libraries to deliver), the app needs less libraries (startup, memory, runtime) and the internal handling of SVG vector data is completely vector-graphic oriented. There were also ODF-compatible File Format adaptions needed, more concrete the in ODF already contained and described multi-image support. In ODF, the original SVG is now embedded to the 'Pictures' folder inside the ODF file as one would expect from such a feature and can be easily extracted (unzip the ODF file and there you are). There is also a Png file written as replacement image. The draw:frame is now multi-image capable (as the spec allows). In the case of a SVG it writes a good quality Png and the original SVG as draw:image elements. Since older (and other) office versions are only capable of loading a single (and thus the first) image, the Png is written first. This allows file exchange with other and older offices without breaking backward compatibility and/or ODF file exchange. At load time, multi-image support will choose the best quality graphic available for further work, e.g. preferring vector format over pixel format, pixel format with alpha over non-alpha and lossless formats over those with losing info (you get the idea). Other ODF implementations (e.g. a viewer) may just use the pixel graphic available. Multi-image support is independent from SVG in principle and will work with all image file formats. This is implemented for the Drawinglayer graphic object (used in Draw/Impress/Calc) and the Writer graphic object (used in Writer). SVG is no longer interpreted each time it needs to be rendered (unavoidable by an external renderer), but only once transformed to a sequence of primitives (UNO API graphic atoms). That sequence is then used for all outputs, transformed to the graphic object's form and viewport. The sequence itself is completely view-independent. Internally, it is reused and thus it makes no difference if you have your SVG graphic added once or multiple times to your document. This is also true for saving, so always only one copy of your added SVG will be written (the same is true for the replacement Png image). Both, the sequence of primitives and the replacement image, are created using new UNO API services. One is capable of converting an io::XInputStream with SVG content to a sequence of primitives, the other is able to convert every sequence of primitives to a rendering::XBitmap with given DPI and discrete sizes (pixels, with automatic resolution reduction to a given maximum square pixel count to be on the safe side). This will be useful for other purposes, too, since it creates a fully alpha-capable representation of anything in primitive format to use as e.g. sprite. For all graphic processing the created vector graphic in form of a sequence of primitives is used. This means that you will get best quality in all zoom situations and all resolutions. This is also true for all exports, e.g. printing or PDF export which also uses the vector format. With an external renderer, it is unavoidable to use bitmaps with discrete solution in those cases, looking bad when zooming and needing more space in most cases as vector data. There is one caveat since not all program paths already use primitives; some will use the internal MetaFile format in-between (One more reason for more reworks to primitive usages in the future). I implemented most SVG features from SVG 1.1, but not yet using animations or interactions (but possible in the future due to an own interpreter, impossible with an external SVG renderer). It supports all geometric SVG forms. It supports SVG gradients (using a new primitive for this which will be reused when we add SVG gradients to SdrObjects one day), these have a resolution-dependent low-level format to not waste runtime on low resolutions. It supports masks, clipPath, markers, linked content, embedded graphics or SVG (intern, extern, base64), SVG use nodes, text, text on curve and patterns. It does not yet support filters, color profiles, embedded scripts, interactions and linking. These can be added when needed, most of them will need to implement new primitive types (e.g. filtering) which would be useful for the future anyways. Especially interesting is the possibility to later add SVG animation import to GraphicObjects for Impress. Some side effects: I had to fix cropping (unified with new primitive) which works now also for mirrored graphics (never worked) and quite some other stuff. We are prepared for SVG gradients as possible future feature (we can already render them now). You can work with an added SVG as with a normal GraphicObject; crop it, break it (to SdrObjects, currently limited to the transfer over the old MetaFile format, though). You can convert an inserted Tux to 3D, you can bend the SVG in vector quality in Draw. It is possible to directly export the original SVG again by selecting the object and using 'Save as Picture...' from the context menu. You can add text, line style, fill style, pretty much the same as most other graphic objects. You can add shadow which casts shadow for the SVG graphic itself as expected (also not possible with an external renderer). This is a bigger change, but most stuff is isolated in the two mentioned services. There will be errors (I'm too long a programmer to deny that :-)), but I tried to be as careful as possible. I already got some help from other community members and fixed some reported bugs (kudos to all testers and bug writers), but to find the rest, your help is appreciated. Please feel free to play around with any SVG you can find in current AOO 3.4 builds and report problems early in the Apache bugtracker! Here is another blog entry about an early version of this feature.And here are some developer snapshots of AOO 3.4 when you want to check it out. Be aware that these are AOO 3.4 Unofficial Developer Snapshots; these are intended to be used for early testing by other community volunteers. They have no release quality and should not be installed in a production environment. Developer snapshots can be unstable and are expected to have bugs. Regards,        Armin [Less]
Posted over 12 years ago by cos
Hadoop is taking central stage in the discussions about processing of the large amount of unstructured data. With raising the popularity of the system I found that people are really puzzled with all the multiplicity of Hadoop versions; the ... [More] small, yet annoying differences introduced by different vendors; the frustration when vendors are trying to lock up their customers using readily available open source data analytic components on top of Hadoop, and on and on. So, after explaining who was born from whom for the 3rd time - and I tell you, drawing neat pictures on a napkin in a coffee shop isn't my favorite activity - I put together this little diagram below. Click on it to inspect it in greater details. A warning: the diagram only includes more or less significant releases of Hadoop and Hadoop-derived systems available today. I don't want to waste any time on some obscure releases or branches which never been accepted at any significant level. The only exception is 0.21 which was a natural continuation of 0.20 and predecessor of recently released 0.22. Some explanations for the diagram: Green rectangles designate official Apache Hadoop releases openly available for anyone in the world for free Black ovals show Hadoop branches that are not yet officially released by Apache Hadoop (or might not be released ever). However, they are usually available in the form of source code or tar-ball artifacts Red ovals are for commercial Hadoop derivatives which might be based on Hadoop or use Hadoop as a part of custom systems (like in case of MapR). These derivatives can be or can be not compatible with Hadoop and Hadoop data processing stack. Once you're presented with the view like this it is getting clear that there are two centers of the gravity in today's universe of elephants: 0.20.2 based releases and derivatives; and 0.22 based branches, future releases, and derivatives. Also, it becomes quite clear which are likely to be sucked into a black hole. The transition from 0.20+ to 0.2[1,2] was real critical because of introduced true HDFS append, fault injection, and code injection for system testing. And the fact that 0.21 hasn't been released for a long time, creating an empty space in the high demand environment. Even after it did come out, it didn't get any traction in the community. Meanwhile, HDFS append was very critical for HBase to move forward, so 0.20.2-append has been created to support the effort. A quite similar story had happened to 0.22: two different release managers was trying to get it out: first gave up, but the second has actually succeeded in pulling an effort of a part of the community towards it. As you can see, HDFS append wasn't available in an official Apache Hadoop release for some time (except for 0.21 with the earlier disclaimer). Eventually it has been merged into 0.20.205 (recently dubbed as Hadoop 1.0) and that allows HBase to be nicely integrated with the official Apache Hadoop without any custom patching process. The release of 0.20.203 was quite significant because it provided a heavily tested Hadoop security, developed by Yahoo! Hadoop development team (known as HortonWorks nowadays). Bits and pieces of 0.20.203 - even before the official release - were absorbed by at least one commercial vendor to add corporate grade Kerberos security to their derivatives of Hadoop (as in case of Cloudera CDH3). The diagram above clearly shows a few important gaps of the rest of commercial offerings: none of them supports Kerberos security (EMC, IBM, and MapR) unavailability of Hbase due to the lack of HDFS append in their systems (EMC, IBM). In case of MapR you end up using a custom HBase distributed by MapR. I don't want to make any speculation of the latter in this article. Apparently, the vacuum of significant releases between 0.20 and 0.22 appeared to be a major urge for Hadoop PMC and now - just days after release of 1.0 - 0.22 got out. With 0.23 already going through release process, championed by HortonWorks team. That release brings in some interesting innovations like Federations and MapReduce 2.0. Once current alpha 0.23 (which might become Hadoop 2.0 or even Hadoop 3.0) is ready for the final release I would expect new versions of commercial distributions springing to live as it was the case before. At this point I will update the diagram :) If you can imagine the variety of the other animals such as Pig, and Hive piling on top of Hadoop, you would get astonished by the complexity of inter-component relations and, more importantly, about intricacies of building a stable data processing stack. This is why project BigTop has been so important and popular ever since it sprung to life last year. Here you can read about Bigtop's relation to Hadoop stack here. Cross-posted [Less]
Posted over 12 years ago by Peter Elst
This is the beginning of a new era for Flex. On December 31st 2011, our project was accepted as an incubator podling at the Apache Software Foundation. Now the real work begins. If you’ve joined the developer mailing list (or checked the archives) ... [More] you will see some of the excitement about Apache Flex and discussions on how to move the technology forward. One thing we can say for sure: we have a good group of enthusiastic volunteers and no lack of ideas. We've seen some great first contributions to the whiteboard - an area in the repository for experimental work - by several committers that could potentially make it in in a next release. Our contributors from Adobe are working with Adobe’s legal department to get the Flex SDK source code submitted to the Apache repository. We expect the initial commit of code will be made by the end of next week. Special thanks to Adobe’s Alex Harui and Carol Frampton for their hard work! How can you get involved? If you're a Flex developer and would like to follow what is happening, you should join the developer mailing list. The mantra at Apache is "if it didn't happen on the mailing list, it didn't happen." We have a JIRA issue tracker being set up, too. This will be an excellent entry point for developers to start contributing code patches. We're also looking for a logo for Apache Flex. If you have good design skills and would like to submit a proposal you have until January 17th. In short, we're excited to see things getting up and running and the amount of enthusiasm shown by the community about the future of Flex at Apache! [Less]
Posted over 12 years ago by arvind
Apache Sqoop (incubating) was created to efficiently transfer bulk data between Hadoop and external structured datastores, such as RDBMS and data warehouses, because databases are not easily accessible by Hadoop. Sqoop is currently undergoing ... [More] incubation at The Apache Software Foundation. More information on this project can be found at http://incubator.apache.org/sqoop. The popularity of Sqoop in enterprise systems confirms that Sqoop does bulk transfer admirably. That said, to enhance its functionality, Sqoop needs to fulfill data integration use-cases as well as become easier to manage and operate. What is Sqoop?As described in a previous blog post, Sqoop is a bulk data transfer tool that allows easy import/export of data from structured datastores such as relational databases, enterprise data warehouses, and NoSQL systems. Using Sqoop, you can provision the data from an external system into HDFS, as well as populate tables in Hive and HBase. Similarly, Sqoop integrates with the workflow coordinator Apache Oozie (incubating), allowing you to schedule and automate import/export tasks. Sqoop uses a connector-based architecture which supports plugins that provide connectivity to additional external systems. Sqoop's ChallengesSqoop has enjoyed enterprise adoption, and our experiences have exposed some recurring ease-of-use challenges, extensibility limitations, and security concerns that are difficult to support in the original design:- Cryptic and contextual command line arguments can lead to error-prone connector matching, resulting in user errors- Due to tight coupling between data transfer and the serialization format, some connectors may support a certain data format that others don't (e.g. direct MySQL connector can't support sequence files)- There are security concerns with openly shared credentials- By requiring root privileges, local configuration and installation are not easy to manage- Debugging the map job is limited to turning on the verbose flag- Connectors are forced to follow the JDBC model and are required to use common JDBC vocabulary (URL, database, table, etc), regardless if it is applicable These challenges have motivated the design of Sqoop 2, which is the subject of this post. That said, Sqoop 2 is a work in progress whose design is subject to change. Sqoop 2 will continue its strong support for command line interaction, while adding a web-based GUI that exposes a simple user interface. Using this interface, a user can walk through an import/export setup via UI cues that eliminate redundant options. Various connectors are added in the application in one place and the user is not tasked with installing or configuring connectors in their own sandbox. These connectors expose their necessary options to the Sqoop framework which then translates them to the UI. The UI is built on top of a REST API that can be used by a command line client exposing similar functionality. The introduction of Admin and Operator roles in Sqoop 2 will restrict 'create' access for Connections to Admins and 'execute' access to Operators. This model will allow integration with platform security and restrict the end user view to only operations applicable to end users. Ease of UseWhereas Sqoop requires client-side installation and configuration, Sqoop 2 will be installed and configured server-side. This means that connectors will be configured in one place, managed by the Admin role and run by the Operator role. Likewise, JDBC drivers will be in one place and database connectivity will only be needed on the server. Sqoop 2 will be a web-based service: front-ended by a Command Line Interface (CLI) and browser and back-ended by a metadata repository. Moreover, Sqoop 2's service level integration with Hive and HBase will be on the server-side. Oozie will manage Sqoop tasks through the REST API. This decouples Sqoop internals from Oozie, i.e. if you install a new Sqoop connector then you won't need to install it in Oozie also. Ease of ExtensionIn Sqoop 2, connectors will no longer be restricted to the JDBC model, but can rather define their own vocabulary, e.g. Couchbase no longer needs to specify a table name, only to overload it as a backfill or dump operation. Common functionality will be abstracted out of connectors, holding them responsible only for data transport. The reduce phase will implement common functionality, ensuring that connectors benefit from future development of functionality. Sqoop 2's interactive web-based UI will walk users through import/export setup, eliminating redundant steps and omitting incorrect options. Connectors will be added in one place, with the connectors exposing necessary options to the Sqoop framework. Thus, users will only need to provide information relevant to their use-case. With the user making an explicit connector choice in Sqoop 2, it will be less error-prone and more predictable. In the same way, the user will not need to be aware of the functionality of all connectors. As a result, connectors no longer need to provide downstream functionality, transformations, and integration with other systems. Hence, the connector developer no longer has the burden of understanding all the features that Sqoop supports. SecurityCurrently, Sqoop operates as the user that runs the 'sqoop' command. The security principal used by a Sqoop job is determined by what credentials the users have when they launch Sqoop. Going forward, Sqoop 2 will operate as a server based application with support for securing access to external systems via role-based access to Connection objects. For additional security, Sqoop 2 will no longer allow code generation, require direct access to Hive and HBase, nor open up access to all clients to execute jobs. Sqoop 2 will introduce Connections as First-Class Objects. Connections, which will encompass credentials, will be created once and then used many times for various import/export jobs. Connections will be created by the Admin and used by the Operator, thus preventing credential abuse by the end user. Furthermore, Connections can be restricted based on operation (import/export). By limiting the total number of physical Connections open at one time and with an option to disable Connections, resources can be managed. SummaryAs detailed in this presentation, Sqoop 2 will enable users to use Sqoop effectively with a minimal understanding of its details by having a web-application run Sqoop, which allows Sqoop to be installed once and used from anywhere. In addition, having a REST API for operation and management will help Sqoop integrate better with external systems such as Oozie. Also, introducing a reduce phase allows connectors to be focused only on connectivity and ensures that Sqoop functionality is uniformly available for all connectors. This facilitates ease of development of connectors. We encourage you to participate in and contribute to Sqoop 2's Design and Development (SQOOP-365). Guest post by Kathleen Ting. [Less]
Posted over 12 years ago by robweir
After 4 years of existence, the Community Forum has moved on the Apache Software Foundation (ASF) servers at the end of October 2011 (see details). Here are some figures about how we are doing on the English forum. We will try to make this kind of ... [More] report on a monthly basis in the forum and perhaps quarterly on the blog. Please remember that the forum is managed by a group of users helping other users for free and on their spare time. Some basics: The number of posts, members and topics is taken from the phpBB information bottom of main index page Solved topics are counted in all the forums except admin and archives sections (not visible to standard users) Note that the ratio solved topics vs. total topics is slightly biased since the topics in the archives and admin sections are counted (in phpBB statistics) but not the solved ones (custom search). However, there are less than 550 topics there (in more than a 40,000 grand total, so less than 1.5%). These last figures shouldn't change very much since the private sections are not very active. Here it is (since the solved topics is a new metric, there is only one point for the moment). The blue line (number of posts) is not that important, it just shows that the trend is consistent with the other metrics.  The most interesting statistics are the red lines for the members and the topics (with triangles, giving the time when the figures have been recorded). Note that the ratio topics vs. members is 0.9 for the English forum and above 1.5 for the French and Japanese forums. We tend not to be too harsh for the rules on the English section and topics are not often split when different users ask questions (still related of course) in the same thread. Activity has a slightly higher slope during the first 2 years. It may be linked to the building of the knowledge database. Once the main issues and common questions have been discussed, users find their answers more and more easily with a mere search, hence less topics needed. Neither the release of LibreOffice (Oct 2010), nor the move to the ASF servers (Oct 2011) have changed anything for the activity. The decrease in the number of members (end of 2011) is linked to the cleaning of banned users. They had just been banned until now but to keep only the "real users", their account has been deleted (nearly 1800). This was the first cleaning ever done from the launch of this forum 4 years ago. 1550 of them had been identified as spammers because of their post(s). 250 were passive spam (link in signature or interest field of the profile, without any post). The ratio of the solved topics is rather good: 14,000 solved (green triangle) in 41,000 (red triangles), that makes more than 1 in 3 (all users don't bother to tag their topic as solved). For the record, last quarter has been in line with the rest: 2100 new users from Oct 1st to end of this year, meaning 1800 new topics. As for the spam, we have had 1800 users in 1500 days, it makes 1.21 spammer a day, still rather low, thanks to the registration process (hard to cheat for bots). Last figures: 2011 has shown an increase in the max number of online users along the months. Peak reached 232 beginning of October. The counter has been reset on Jan 1st 2012 and is already at 214, proving the audience is still there. Some words about the team. Let's not forget Terry Ellison who was the main maintainer of the forum until the move to Apache Software Foundation servers. His huge involvement has made it possible for both a clean running of the forum during 4 years, making it a great place for those needing/providing help, and contributing to the transfer of the forum to the ASF servers with a minimal impact for the users. Hagar Delest, On the behalf of the Forum Volunteers [Less]
Posted over 12 years ago by alg
I just wanted to send some notes about added features which are part of AOO3.4 version. This one is actually the result of fixing tasks #118558#, #118485#, #108221# and #67705# which are all about GraphicObjects, OLEObjects (OLE means Object Linking ... [More] and Embedding) and their geometrical attributes and properties. You may take a look at the tasks if you are interested in details, here I want to describe the benefits. GraphicObjects are used when you insert a picture (pixel and vector data) or convert something to it. They already supported the full attribute set, so line style, fill style, text and shadow are possible. Geometrically, they could be transformed widely, but could not be sheared. Because now the content of GraphicObjects is displayed using primitives (and these are fully transformable) it is possible to also use shear and thus now completely support all geometrical transformations used in the office. More interesting is that this is also true for OLEObjects, thus I added all these possibilities to OLEObjects of any kind, not only to our own internal OLEObjects (e.g. Chart, mathematical formula), but all possible external OLEObjects. These can now have line styles, fill styles, text and shadow and can be fully transformed. It is also possible to convert them to GraphicObjects which is the base for converting them to something else. Thus, you may now slant or distort OLEObjects, convert them to GraphicObjects, make geometrical modifications like merge/subtract/intersect with other objects or even convert them to 3D objects.Some of the possible changes may lead too far for daily use, but some are pretty useful. It is now e.g. possible to add a mathematical formula and position it somewhere vertically by rotating it 90 degrees. It is possible to rotate chart OLEObjects themselves to get chart displays which the chart itself does not directly support. It is also useful to add a frame to a chart. With using the text offset it is also possible e.g. to add text to the OLEObject and move the text outside the object to get a caption. I leave more possibilities to your imagination... Some examples: (a) Math OLEObject rotated 90 degrees, blue filled and with border(b) Chart OLEObject with gradient fill, border and object text as caption(c) Same chart without fill, 90 degree rotation and shadow(d) Chart bend in 2D, converted to GraphicObject (no longer an OLEObject)(e) Chart OLEObject converted to 3D (only for demonstration, not too nice to view...)(f) The math OLEObject, converted and bent, fill removed I hope you got an impression; I'm not the designer guy, so excuse the examples. Regards,      Armin [Less]
Posted over 12 years ago by Sally
Open Source "Big Data" Cloud computing platform powers millions of compute-hours to process exabytes of data for Amazon.com, AOL, Apple, eBay, Facebook, foursquare, HP, IBM, LinkedIn, Microsoft, Netflix, The New York Times, Rackspace, Twitter ... [More] , Yahoo!, and more. 4 January 2012 —FOREST HILL, MD— The Apache Software Foundation (ASF), the all-volunteer developers, stewards, and incubators of nearly 150 Open Source projects and initiatives, today announced Apache™ Hadoop™ v1.0, the Open Source software framework for reliable, scalable, distributed computing. The project’s latest release marks a major milestone six years in the making, and has achieved the level of stability and enterprise-readiness to earn the 1.0 designation. A foundation of Cloud computing and at the epicenter of "big data" solutions, Apache Hadoop enables data-intensive distributed applications to work with thousands of nodes and exabytes of data. Hadoop enables organizations to more efficiently and cost-effectively store, process, manage and analyze the growing volumes of data being created and collected every day. Apache Hadoop connects thousands of servers to process and analyze data at supercomputing speed.  "This release is the culmination of a lot of hard work and cooperation from a vibrant Apache community group of dedicated software developers and committers that has brought new levels of stability and production expertise to the Hadoop project," said Arun C. Murthy, Vice President of Apache Hadoop. "Hadoop is becoming the de facto data platform that enables organizations to store, process and query vast torrents of data, and the new release represents an important step forward in performance, stability and security. "Originating with technologies developed by Yahoo, Google, and other Web 2.0 pioneers in the mid-2000s, Hadoop is now central to the big data strategies of enterprises, service providers, and other organizations," wrote James Kobielus in the independent Forrester Research, Inc. report, "Enterprise Hadoop: The Emerging Core Of Big Data" (October 2011). Dubbed a "Swiss army knife of the 21st century" and named "Innovation of the Year" by the 2011 Media Guardian Innovation Awards, Apache Hadoop is widely deployed at organizations around the globe, including industry leaders from across the Internet and social networking landscape such as Amazon Web Services, AOL, Apple, eBay, Facebook, foursquare, HP, LinkedIn, Netflix, The New York Times, Rackspace, Twitter, and Yahoo!. Other technology leaders such as Microsoft and IBM have integrated Apache Hadoop into their offerings. Yahoo!, an early pioneer, hosts the world’s largest known Hadoop production environment to date, spanning more than 42,000 nodes. "Achieving the 1.0 release status is a momentous achievement from the Apache Hadoop community and the result of hard development work and shared learnings over the years," said Jay Rossiter, senior vice president, Cloud Platform Group at Yahoo!. "Apache Hadoop will continue to be an important area of investment for Yahoo!. Today Hadoop powers every click at Yahoo!, helping to deliver personalized content and experiences to more than 700 million consumers worldwide." "Apache Hadoop is in use worldwide in many of the biggest and most innovative data applications," said Eric Baldeschwieler, CEO of Hortonworks. "The v1.0 release combines proven scalability and reliability with security and other features that make Apache Hadoop truly enterprise-ready." "Gartner is seeing a steady increase in interest in Apache Hadoop and related "big data" technologies, as measured by substantial growth in client inquiries, dramatic rises in attendance at industry events, increasing financial investments and the introduction of products from leading data management and data integration software vendors," said Merv Adrian, Research Vice President at Gartner, Inc. "The 1.0 release of Apache Hadoop marks a major milestone for this open source offering as enterprises across multiple industries begin to integrate it into their technology architecture plans." Apache Hadoop v1.0 reflects six years of development, production experience, extensive testing, and feedback from hundreds of knowledgeable users, data scientists, systems engineers, bringing a highly stable, enterprise-ready release of the fastest-growing big data platform. It includes support for: • HBase (sync and flush support for transaction logging) • Security (strong authentication via Kerberos) • Webhdfs (RESTful API to HDFS) • Performance enhanced access to local files for HBase • Other performance enhancements, bug fixes, and features • All version 0.20.205 and prior 0.20.2xx features "We are excited to celebrate Hadoop's milestone achievement," said William Lazzaro, Director of Engineering at Concurrent Computer Corporation. "Implementing Hadoop at Concurrent has enabled us to transform massive amounts of real-time data into actionable business insights, and we continue to look forward to the ever-improving iterations of Hadoop." "Hadoop, the first ubiquitous platform to emerge from the ongoing proliferation of Big Data and noSQL technologies, is set to make the transition from Web to Enterprise technology in 2012," said James Governor, co-founder of RedMonk, "driven by adoption and integration by every major vendor in the commercial data analytics market. The Apache Software Foundation plays a crucial role in supporting the platform and its ecosystem." Availability and Oversight As with all Apache products, Apache Hadoop software is released under the Apache License v2.0, and is overseen by a self-selected team of active contributors to the project. A Project Management Committee (PMC) guides the Project's day-to-day operations, including community development and product releases. Apache Hadoop release notes, source code, documentation, and related resources are available at http://hadoop.apache.org/. About The Apache Software Foundation (ASF) Established in 1999, the all-volunteer Foundation oversees nearly one hundred fifty leading Open Source projects, including Apache HTTP Server — the world's most popular Web server software. Through the ASF's meritocratic process known as "The Apache Way," more than 350 individual Members and 3,000 Committers successfully collaborate to develop freely available enterprise-grade software, benefiting millions of users worldwide: thousands of software solutions are distributed under the Apache License; and the community actively participates in ASF mailing lists, mentoring initiatives, and ApacheCon, the Foundation's official user conference, trainings, and expo. The ASF is a US 501(3)(c) not-for-profit charity, funded by individual donations and corporate sponsors including AMD, Basis Technology, Cloudera, Facebook, Google, IBM, HP, Hortonworks, Matt Mullenweg, Microsoft, PSW Group, SpringSource/VMware, and Yahoo!. For more information, visit http://www.apache.org/. "Apache", "Apache Hadoop", and "ApacheCon" are trademarks of The Apache Software Foundation. All other brands and trademarks are the property of their respective owners. # # # Media Contact:Sally KhudairiVice President, The Apache Software Foundation+1 617 921 8656 <[email protected]> [Less]