Search Results

Search found 23044 results on 922 pages for 'oracle solaris 11'.

Page 496/922 | < Previous Page | 492 493 494 495 496 497 498 499 500 501 502 503  | Next Page >

  • APEX auf der DOAG2013

    - by Carsten Czarski
    Auch dieses Jahr trifft sich die Oracle Community wieder zur DOAG2013, der größten Oracle-Konferenz im deutschsprachigen Raum. Zum Thema APEX gibt es, wie schon in den letzten Jahren, in einem eigenen Stream mit mehr als 20 Vorträgen. Bekannte Sprecher wie Patrick Wolf, Niels de Bruijn, Peter Raganitsch und andere werden vor Ort sein - die DOAG2013 ist also die Gelegenheit zum Kennenlernen, Austausch von Neuigkeiten und zur Diskussion. Das in den letzten Jahren sehr gut angenommene APEX Experten Panel findet auch dieses Jahr wieder statt. Am 21. November um 11:00 beantwortet das Entwicklerteam im Raum Hong Kong Ihre Fragen rund um Application Express. Bitte reichen Sie uns Ihre Fragen vorab auf dieser Webseite ein. Sehen wir uns auf der DOAG2013 ...?

    Read the article

  • Weblogic Virtual Developer Day Reminder - Feb 1 @ 9:30am PT

    - by Cassandra Clark
    Don't forget to register and attend the next Oracle Technology Network Virtual Developer Day- Weblogic Server tomorrow starting at 9:30 am PST. Learn how Oracle WebLogic Server enables a whole new level of productivity for enterprise developers. Also hear the latest on Java EE 6 and the programming tenets that have made it a true platform breakthrough, with new programming paradigms, persistence strategies, and more: * Convention over configuration - minimal XML * Leaner and meaner API - and one that is an open standard * POJO model - managed beans for testable components * Annotation-based programming model - decorate and inject * Reduce or eliminate need for deployment descriptors * Traditional API for advanced users We will have three live events - N. America - Tuesday, Feb 1, 2011 09:30 a.m. - 1:30 p.m. US Pacific TIme Register Europe / Russia - February 10, 2011 9:30 a.m. UK Time / 10:30 a.m. CET Register India - February 17, 2011 9:30 a.m. India time Register

    Read the article

  • JCP wrap up of JavaOne and some news links

    - by heathervc
    A New article has been published on JCP.org covering the happenings surrounding JavaOne earlier this month, JCP Wrap-Up: Platform Evolution and JCP Enthusiasm. The JCP sessions from JavaOne are also now available from the JavaOne conference catalog: JCP.Next: Reinvigorating Java Standards Session ID: BOF6272 101 Ways to Improve Java: Why Developer Participation Matters Session ID: BOF6283 Meet the JCP Executive Committee Candidates Session ID: BOF6307 And a few links from around the web related to the JCP and JavaOne are below. Let us know if we've missed anything! 101 Ways to Improve Java - Why Developer Participation Matters (DZone) JavaOne 2012: 101 Ways to Improve Java (Java Code Geeks) JCP Ready for the Hard Stuff: Revising the JSPA (Application Development Trends) Oracle Hails Community Involvement in Java (v3.co.uk) JavaOne 2012 Recap Trip Report: Oracle OpenWorld and JavaOne 2012

    Read the article

  • Auf Erfolg spezialisiert

    - by A&C Redaktion
    Spezialisierung kommt an. So lautet kurz gefasst die Bilanz, die Oracle Alliances & Channel jetzt, nach gut einem Jahr Laufzeit des OPN Spezialisierungsprogramms, gezogen hat. Das Wichtigste auf einen Blick: über 400 Unternehmen in 65 Ländern in Europa, Afrika und dem Nahen Osten sind bereits spezialisierte Oracle Partner270 davon haben das Platin-Level erreichtinsgesamt erfolgten über 560 Spezialisierungensomit sind 14.400 spezialisierte Fachkräfte im OPN Netzwerk tätig und das in 65 Ländern der EMEA-Region Als Grund dafür dass die Sepzialisierungsangebote so gut ankommen, nennt Stein Surlien, dass sich Partner "besser vom Wettbewerb abheben und für ihr spezifisches Fachwissen anerkannt" werden. Weitere Fakten, Stimmen und Einschützungen finden Sie in der aktuellen Pressemitteilung.

    Read the article

  • Cloud Integration White Paper - Now Available

    - by Bruce Tierney
    Interested in expanding your existing application infrastructure to integrate with cloud applications?  Download the new Oracle White Paper "Cloud Integration - A Comprehensive Solution" to learn not just about connectivity but the other key aspects of successful cloud integration. The paper includes three technical examples of cloud integration with Oracle Fusion Applications, Saleforce, and Workday and follows with the importance of taking a comprehensive approach to also include service aggregation, service virtualization, cloud security considerations and the benefit of maintaining a unified approach to monitoring and management despite an increasingly distributed hybrid infrastructure. To keep the integration architecture from being defined "accidentally" as new business units subscribe to additional cloud vendors outside the participation of IT, a discussion on the "Accidental SOA Cloud Architecture" is included: As shown in the table of contents below, the white paper provides a combination of high-level awareness about key considerations as well as a technical deep dive of the steps needed for cloud integration connectivity: Hope you find the White Paper valuable.  Please download from the following link

    Read the article

  • Azure &ndash; Part 5 &ndash; Repository Pattern for Table Service

    - by Shaun
    In my last post I created a very simple WCF service with the user registration functionality. I created an entity for the user data and a DataContext class which provides some methods for operating the entities such as add, delete, etc. And in the service method I utilized it to add a new entity into the table service. But I didn’t have any validation before registering which is not acceptable in a real project. So in this post I would firstly add some validation before perform the data creation code and show how to use the LINQ for the table service.   LINQ to Table Service Since the table service utilizes ADO.NET Data Service to expose the data and the managed library of ADO.NET Data Service supports LINQ we can use it to deal with the data of the table service. Let me explain with my current example: I would like to ensure that when register a new user the email address should be unique. So I need to check the account entities in the table service before add. If you remembered, in my last post I mentioned that there’s a method in the TableServiceContext class – CreateQuery, which will create a IQueryable instance from a given type of entity. So here I would create a method under my AccountDataContext class to return the IQueryable<Account> which named Load. 1: public class AccountDataContext : TableServiceContext 2: { 3: private CloudStorageAccount _storageAccount; 4:  5: public AccountDataContext(CloudStorageAccount storageAccount) 6: : base(storageAccount.TableEndpoint.AbsoluteUri, storageAccount.Credentials) 7: { 8: _storageAccount = storageAccount; 9:  10: var tableStorage = new CloudTableClient(_storageAccount.TableEndpoint.AbsoluteUri, 11: _storageAccount.Credentials); 12: tableStorage.CreateTableIfNotExist("Account"); 13: } 14:  15: public void Add(Account accountToAdd) 16: { 17: AddObject("Account", accountToAdd); 18: SaveChanges(); 19: } 20:  21: public IQueryable<Account> Load() 22: { 23: return CreateQuery<Account>("Account"); 24: } 25: } The method returns the IQueryable<Account> so that I can perform the LINQ operation on it. And back to my service class, I will use it to implement my validation. 1: public bool Register(string email, string password) 2: { 3: var storageAccount = CloudStorageAccount.FromConfigurationSetting("DataConnectionString"); 4: var accountToAdd = new Account(email, password) { DateCreated = DateTime.Now }; 5: var accountContext = new AccountDataContext(storageAccount); 6:  7: // validation 8: var accountNumber = accountContext.Load() 9: .Where(a => a.Email == accountToAdd.Email) 10: .Count(); 11: if (accountNumber > 0) 12: { 13: throw new ApplicationException(string.Format("Your account {0} had been used.", accountToAdd.Email)); 14: } 15:  16: // create entity 17: try 18: { 19: accountContext.Add(accountToAdd); 20: return true; 21: } 22: catch (Exception ex) 23: { 24: Trace.TraceInformation(ex.ToString()); 25: } 26: return false; 27: } I used the Load method to retrieve the IQueryable<Account> and use Where method to find the accounts those email address are the same as the one is being registered. If it has I through an exception back to the client side. Let’s run it and test from my simple client application. Oops! Looks like we encountered an unexpected exception. It said the “Count” is not support by the ADO.NET Data Service LINQ managed library. That is because the table storage managed library (aka. TableServiceContext) is based on the ADO.NET Data Service and it supports very limit LINQ operation. Although I didn’t find a full list or documentation about which LINQ methods it supports I could even refer a page on msdn here. It gives us a roughly summary of which query operation the ADO.NET Data Service managed library supports and which doesn't. As you see the Count method is not in the supported list. Not only the query operation, there inner lambda expression in the Where method are limited when using the ADO.NET Data Service managed library as well. For example if you added (a => !a.DateDeleted.HasValue) in the Where method to exclude those deleted account it will raised an exception said "Invalid Input". Based on my experience you should always use the simple comparison (such as ==, >, <=, etc.) on the simple members (such as string, integer, etc.) and do not use any shortcut methods (such as string.Compare, string.IsNullOrEmpty etc.). 1: // validation 2: var accountNumber = accountContext.Load() 3: .Where(a => a.Email == accountToAdd.Email) 4: .ToList() 5: .Count; 6: if (accountNumber > 0) 7: { 8: throw new ApplicationException(string.Format("Your account {0} had been used.", accountToAdd.Email)); 9: } We changed the a bit and try again. Since I had created an account with my mail address so this time it gave me an exception said that the email had been used, which is correct.   Repository Pattern for Table Service The AccountDataContext takes the responsibility to save and load the account entity but only for that specific entity. Is that possible to have a dynamic or generic DataContext class which can operate any kinds of entity in my system? Of course yes. Although there's no typical database in table service we can threat the entities as the records, similar with the data entities if we used OR Mapping. As we can use some patterns for ORM architecture here we should be able to adopt the one of them - Repository Pattern in this example. We know that the base class - TableServiceContext provide 4 methods for operating the table entities which are CreateQuery, AddObject, UpdateObject and DeleteObject. And we can create a relationship between the enmity class, the table container name and entity set name. So it's really simple to have a generic base class for any kinds of entities. Let's rename the AccountDataContext to DynamicDataContext and make the type of Account as a type parameter if it. 1: public class DynamicDataContext<T> : TableServiceContext where T : TableServiceEntity 2: { 3: private CloudStorageAccount _storageAccount; 4: private string _entitySetName; 5:  6: public DynamicDataContext(CloudStorageAccount storageAccount) 7: : base(storageAccount.TableEndpoint.AbsoluteUri, storageAccount.Credentials) 8: { 9: _storageAccount = storageAccount; 10: _entitySetName = typeof(T).Name; 11:  12: var tableStorage = new CloudTableClient(_storageAccount.TableEndpoint.AbsoluteUri, 13: _storageAccount.Credentials); 14: tableStorage.CreateTableIfNotExist(_entitySetName); 15: } 16:  17: public void Add(T entityToAdd) 18: { 19: AddObject(_entitySetName, entityToAdd); 20: SaveChanges(); 21: } 22:  23: public void Update(T entityToUpdate) 24: { 25: UpdateObject(entityToUpdate); 26: SaveChanges(); 27: } 28:  29: public void Delete(T entityToDelete) 30: { 31: DeleteObject(entityToDelete); 32: SaveChanges(); 33: } 34:  35: public IQueryable<T> Load() 36: { 37: return CreateQuery<T>(_entitySetName); 38: } 39: } I saved the name of the entity type when constructed for performance matter. The table name, entity set name would be the same as the name of the entity class. The Load method returned a generic IQueryable instance which supports the lazy load feature. Then in my service class I changed the AccountDataContext to DynamicDataContext and that's all. 1: var accountContext = new DynamicDataContext<Account>(storageAccount); Run it again and register another account. The DynamicDataContext now can be used for any entities. For example, I would like the account has a list of notes which contains 3 custom properties: Account Email, Title and Content. We create the note entity class. 1: public class Note : TableServiceEntity 2: { 3: public string AccountEmail { get; set; } 4: public string Title { get; set; } 5: public string Content { get; set; } 6: public DateTime DateCreated { get; set; } 7: public DateTime? DateDeleted { get; set; } 8:  9: public Note() 10: : base() 11: { 12: } 13:  14: public Note(string email) 15: : base(email, string.Format("{0}_{1}", email, Guid.NewGuid().ToString())) 16: { 17: AccountEmail = email; 18: } 19: } And no need to tweak the DynamicDataContext we can directly go to the service class to implement the logic. Notice here I utilized two DynamicDataContext instances with the different type parameters: Note and Account. 1: public class NoteService : INoteService 2: { 3: public void Create(string email, string title, string content) 4: { 5: var storageAccount = CloudStorageAccount.FromConfigurationSetting("DataConnectionString"); 6: var accountContext = new DynamicDataContext<Account>(storageAccount); 7: var noteContext = new DynamicDataContext<Note>(storageAccount); 8:  9: // validate - email must be existed 10: var accounts = accountContext.Load() 11: .Where(a => a.Email == email) 12: .ToList() 13: .Count; 14: if (accounts <= 0) 15: throw new ApplicationException(string.Format("The account {0} does not exsit in the system please register and try again.", email)); 16:  17: // save the note 18: var noteToAdd = new Note(email) { Title = title, Content = content, DateCreated = DateTime.Now }; 19: noteContext.Add(noteToAdd); 20: } 21: } And updated our client application to test the service. I didn't implement any list service to show all notes but we can have a look on the local SQL database if we ran it at local development fabric.   Summary In this post I explained a bit about the limited LINQ support for the table service. And then I demonstrated about how to use the repository pattern in the table service data access layer and make the DataContext dynamically. The DynamicDataContext I created in this post is just a prototype. In fact we should create the relevant interface to make it testable and for better structure we'd better separate the DataContext classes for each individual kind of entity. So it should have IDataContextBase<T>, DataContextBase<T> and for each entity we would have class AccountDataContext<Account> : IDataContextBase<Account>, DataContextBase<Account> { … } class NoteDataContext<Note> : IDataContextBase<Note>, DataContextBase<Note> { … }   Besides the structured data saving and loading, another common scenario would be saving and loading some binary data such as images, files. In my next post I will show how to use the Blob Service to store the bindery data - make the account be able to upload their logo in my example.   Hope this helps, Shaun   All documents and related graphics, codes are provided "AS IS" without warranty of any kind. Copyright © Shaun Ziyan Xu. This work is licensed under the Creative Commons License.

    Read the article

  • Sang Shin is back!

    - by arungupta
    Sang Shin is back at Oracle. He is well known for his wide range of Java tutorials at javapassion.com. As a former Sun evangelist, he helped GlassFish and Java EE adoption and continued to do even after his days at Sun. In his current role, he will help evangelize and drive adoption of the Java EE platform, GlassFish, and WebLogic server. You will see him online, speaking at conferences, blogging, and at our different social media channels. Say hello to him whenever you meet him. Here is what Sang had to say about his comeback: I AM BACK to where it all started. It sure felt like coming back home and it's good to see the first hand the spirit of innovation still is alive and well here in Oracle especially on Java front. Looking forward to be engaged with everyone in Java EE community with Passion again! Looking forward to working with Sang! You can ping him at @javapassion.

    Read the article

  • Twitter Tuesday - Top 10 @ArchBeat Tweets - May 20-26, 2014

    - by OTN ArchBeat
    What's everyone looking at? The list below represents the Top 10 most popular tweets for the last seven  days (May 20-26, 2014) among 2,845 people now following @OTNArchBeat. Video: #KScope14 Preview: @stewartbryson talks OBIEE, ODI, and GoldenGate @ODTUG #oracleace May 21, 2014 at 12:00 AM May edition of Oracle's Architect Community newsletter. Features on #WebLogic #WebCenter #SOA #Cloud. May 21, 2014 at 12:00 AM Oracle #ADF and Simplified UI Apps: I18n Feng Shui on Display | @Ultan May 22, 2014 at 12:00 AM The OTNArchBeat Daily is out! Stories via @JavaOneConf @arungupta May 20, 2014 at 12:00 AM Video: #WebLogic Server Templates | @FrankMunz May 21, 2014 at 12:00 AM Supporting multiple #SOASuite revisions with Edition-Based Redefinition | Betty van Dongen May 21, 2014 at 12:00 AM The OTNArchBeat Daily is out! Stories via @soacommunity @oraclebase @InfoQ May 24, 2014 at 12:00 AM Development Lifecycle for Task Flows in #WebCenter Portal | Lyudmil Pelov May 20, 2014 at 12:00 AM Manos libres y vista al frente: Con el futuro puesto #wearables May 21, 2014 at 12:00 AM #GoldenGate: Understanding OGG-01161 Bad Column Index Error | Loren Penton May 21, 2014 at 12:00 AM

    Read the article

  • The convergence of Risk and Performance Management

    Historically, the market has viewed Enterprise Performance Management (EPM) and Governance, Risk and Compliance (GRC) as separate processes and solutions. But these two worlds are coming together – in fact industry analyst firms such as AMR Research believe that by the end of 2009, risk management will be part of every EPM discussion. Tune into this conversation with John O'Rourke, VP of Product Marketing for Oracle Enterprise Performance Management Solutions, and Karen dela Torre, Senior Director of Product Marketing for Financial Applications to learn how EPM and GRC are converging, what the integration points are, and what Oracle is doing to help customers perform more effective risk and performance management.

    Read the article

  • OpenWorld in Small Bites

    - by Kathryn Perry
    Fifty thousand attendees -- that's bigger than the cities some of us live in. Monday morning it took 20 minutes to get from Hall D in Moscone North to a conference room in Moscone South -- the crowds were crushing! A great start to a great week! Larry is as big a name as ever on the program schedule and on the Moscone stage. People were packed in Hall D and clustered around every big screen TV. He stayed on script as he laid out Oracle's SaaS, PaaS, and IaaS strategies. Every seat in Chris Leone's Fusion Apps Cloud Overview was filled on Monday morning. Oracle employees who wanted to get in were turned away. And the same thing happened in the repeat session on Wednesday. Our newest suite of apps is hot! Speaking of hot, the weather was made to order. Then it turned very San Francisco-like on Wednesday afternoon. Downright cold for those who trusted SF temps to hold in the 80's. Who did you follow on Twitter during the conference? So many voices, opinions, and convos! Great combo of social media and sharp minds. Be sure to follow @larryellison, @stevenrmiranda, and @Oracle for updates and MyPOVs. Keywords for the Apps customers at the conference were cloud, mobile, and social. Every day, every session, every speaker. Wednesday afternoon, 4 pm at the Four Seasons hotel. A large roomful of analysts and influencers firing questions at a panel of eight Fusion customers. Steve Miranda moderating. Good energy and a great exchange of information and confidence. Word on the street is that OpenWorld has outgrown San Francisco -- but moving it seems unthinkable. The city isn't just a backdrop for an industry conference - it's a headliner right up there with Larry Ellison and Pearl Jam. As you can imagine, electrical outlets were in high demand at every venue. The most popular hotels and bars near Moscone designed their interiors around accessible electrical power strips. People are plenty willing to buy a drink while they grab a charge. Wednesday afternoon, 4 pm at the Four Seasons hotel. A large roomful of analysts and influencers firing questions at a panel of eight Fusion customers. Steve Miranda moderating. Good energy and a great exchange of information and confidence. Treasure Island in the dark. Eddy Vedder has an amazing voice! And Kings of Leon over delivered on people's expectations. It was cold. It was windy. It was very fun. One analyst said it's the best customer appreciation party in the industry. 

    Read the article

  • APEX Patchset 4.2.1 erschienen

    - by Carsten Czarski
    Seit Freitag, dem 14.12. steht das APEX Patchset 4.2.1 zur Verfügung. Neben zahlreichen Bugfixes sind auch einige, kleinere Änderungen enthalten: Die JQuery Mobile Bibliothek wurde auf Version 1.2.0 aktualisiert Die AnyChart-Engine wurde auf Version 6.0.11 gehoben. Dadurch stehen "Circular Gauge" Charts auch als HTML5 Diagramme bereit Diagramme in Anwendungen für mobile Endgeräte können sich nun dynamisch der Bildschirmgröße anpassen Details zum Patchset finden sich, wie immer in den Release Notes. Wie bislang unterscheidet sich der Installationsvorgang je nach verwendeter APEX-Version. Wenn noch kein APEX oder eine ältere Version als 4.2 installiert ist, kann die Vollinstallation für APEX 4.2.1 direkt aus dem OTN heruntergeladen und normal installiert werden. Wenn APEX 4.2.0 installiert ist, muss Patch Nr. 14732511 aus MyOracle Support heruntergeladen und eingespielt werden.

    Read the article

  • Upload Recordings Of The Problem To Your SR

    - by bseckin
    Do you find yourself trying several times to explain a problem in a Service Request? Does the support engineer ask more than once for clarification? If so, you might be interested in DITO -- Demo It To Oracle. DITO uses CamStudio (free download!) to record the exact nature of the problem, and upload the output to your SR. The following articles provide more details: Working with Support - MOSSOS (Doc ID 1265130.1) "Demo It To Oracle" (DITO) - CamStudio Help (Doc ID 11.1) Why take up valuable time first explaining the problem, then trying to get a web conference setup to show exactly what is going on? The next time you file an SR, try including a recording showing exactly which application is failing, where it is failing, and what it looks like when it fails.

    Read the article

  • Increase productivity, accelerate work-to-cash cycles, and reduce overall firm and client risk with

    Law firms around the world are faced with increasing pressures to do business faster and more efficiently. Learn how firms can automate manual, paper-driven processes, ensure regulatory compliance, integrate systems and offices brought together by mergers and acquisitions, and take on new business quickly and efficiently. Understand how firms can automate manual tasks with Oracle's Whitehill One; get invoices out the door faster with Whitehill Enterprise; and can go green with Whitehill Pre-Bill. In this session, you will hear about Oracle's new legal services offerings that accelerate work-to-cash cycles, increase productivity, and reduce overall firm and client risk.

    Read the article

  • Prosit Neujahr, auf ein schönes und erfolgreiches Jahr 2011!

    - by A&C Redaktion
    Wir hoffen, Sie sind gut im neuen Jahr angekommen und wünschen Ihnen für 2011 alles Gute, Gesundheit und natürlich ein erfolgreiches Geschäftsjahr!Hinter Oracle Alliances & Channels liegt ein Jahr voller Herausforderungen und wir freuen uns darauf, 2011 an die gute und vertrauensvolle Zusammenarbeit mit unseren Partnern anzuknüpfen. Im Januar starten wir gleich mit einem großen Partner-Event ins neue Jahr, demOPN Day Virtual am 24. Januar 2011Von 14 bis 17 Uhr dreht sich dort alles um die Spezialisierung: Von den ersten Schritten, über den Nutzen bis hin zu Erfahrungsberichten einzelner Partner wird das OPN Specialized von allen Seiten beleuchtet. Zum Fragenstellen laden die Q&A-Sessions ein, beliebt ist außerdem die Networking Lounge, in der sich Partner und Interessierte individuell austauschen und vernetzen können.Was sonst noch passiert, im Jahr 2011, das steht laufend aktualisiert im Oracle Channel Event Calender.

    Read the article

  • From Transactions To Engagement

    - by David Dorf
    I've mentioned in the past that Oracle has invested quite a bit in acquiring social companies to build out its Social Relationship Management suite.  The concept is to shift away from transactions and towards engagement.  Social media represents a great opportunity to engage with customers, learn what they want, and personalize the shopping experience for them. I look at SRM as the bridge between traditional CRM and CX.  If you're looking for ideas, check out Five Social Retailing Suggestions and Social Analytics and the Customer.  There are lots of ways to leverage social media to enhance the customer experience and thus drive more sales. My friends over at 8th Bridge have just released their Social IQ report in which they rate retailers on their social capabilities.  They also produced a nice infographic so you can consume the data quickly, but I'd still encourage you to download the full report. Retailers interested in upping their SRM abilities should definitely stop by the Oracle booth at NRF in January.

    Read the article

  • Take Two: Comparing JVMs on ARM/Linux

    - by user12608080
    Although the intent of the previous article, entitled Comparing JVMs on ARM/Linux, was to introduce and highlight the availability of the HotSpot server compiler (referred to as c2) for Java SE-Embedded ARM v7,  it seems, based on feedback, that everyone was more interested in the OpenJDK comparisons to Java SE-E.  In fact there were two main concerns: The fact that the previous article compared Java SE-E 7 against OpenJDK 6 might be construed as an unlevel playing field because version 7 is newer and therefore potentially more optimized. That the generic compiler settings chosen to build the OpenJDK implementations did not put those versions in a particularly favorable light. With those considerations in mind, we'll institute the following changes to this version of the benchmarking: In order to help alleviate an additional concern that there is some sort of benchmark bias, we'll use a different suite, called DaCapo.  Funded and supported by many prestigious organizations, DaCapo's aim is to benchmark real world applications.  Further information about DaCapo can be found at http://dacapobench.org. At the suggestion of Xerxes Ranby, who has been a great help through this entire exercise, a newer Linux distribution will be used to assure that the OpenJDK implementations were built with more optimal compiler settings.  The Linux distribution in this instance is Ubuntu 11.10 Oneiric Ocelot. Having experienced difficulties getting Ubuntu 11.10 to run on the original D2Plug ARMv7 platform, for these benchmarks, we'll switch to an embedded system that has a supported Ubuntu 11.10 release.  That platform is the Freescale i.MX53 Quick Start Board.  It has an ARMv7 Coretex-A8 processor running at 1GHz with 1GB RAM. We'll limit comparisons to 4 JVM implementations: Java SE-E 7 Update 2 c1 compiler (default) Java SE-E 6 Update 30 (c1 compiler is the only option) OpenJDK 6 IcedTea6 1.11pre 6b23~pre11-0ubuntu1.11.10.2 CACAO build 1.1.0pre2 OpenJDK 6 IcedTea6 1.11pre 6b23~pre11-0ubuntu1.11.10.2 JamVM build-1.6.0-devel Certain OpenJDK implementations were eliminated from this round of testing for the simple reason that their performance was not competitive.  The Java SE 7u2 c2 compiler was also removed because although quite respectable, it did not perform as well as the c1 compilers.  Recall that c2 works optimally in long-lived situations.  Many of these benchmarks completed in a relatively short period of time.  To get a feel for where c2 shines, take a look at the first chart in this blog. The first chart that follows includes performance of all benchmark runs on all platforms.  Later on we'll look more at individual tests.  In all runs, smaller means faster.  The DaCapo aficionado may notice that only 10 of the 14 DaCapo tests for this version were executed.  The reason for this is that these 10 tests represent the only ones successfully completed by all 4 JVMs.  Only the Java SE-E 6u30 could successfully run all of the tests.  Both OpenJDK instances not only failed to complete certain tests, but also experienced VM aborts too. One of the first observations that can be made between Java SE-E 6 and 7 is that, for all intents and purposes, they are on par with regards to performance.  While it is a fact that successive Java SE releases add additional optimizations, it is also true that Java SE 7 introduces additional complexity to the Java platform thus balancing out any potential performance gains at this point.  We are still early into Java SE 7.  We would expect further performance enhancements for Java SE-E 7 in future updates. In comparing Java SE-E to OpenJDK performance, among both OpenJDK VMs, Cacao results are respectable in 4 of the 10 tests.  The charts that follow show the individual results of those four tests.  Both Java SE-E versions do win every test and outperform Cacao in the range of 9% to 55%. For the remaining 6 tests, Java SE-E significantly outperforms Cacao in the range of 114% to 311% So it looks like OpenJDK results are mixed for this round of benchmarks.  In some cases, performance looks to have improved.  But in a majority of instances, OpenJDK still lags behind Java SE-Embedded considerably. Time to put on my asbestos suit.  Let the flames begin...

    Read the article

  • Building a Roadmap for an IAM Platform

    - by B Shashikumar
    Identity Management is no longer a departmental solution, it has become a strategic part of every organization's security posture. Enterprises require a forward thinking Identity Management strategy. In our previous blog post on "The Oracle Platform Approach", we discussed a recent study by Aberdeen which showed that organizations taking a platform approach can reduce cost by as much as 48% and have 35% fewer audit deficiencies. So how does an organization get started with an Identity Management (IAM) Platform? What are the components of such a platform and how can an organization continuously evolve it for better ROI and IT agility. What are some of the best practices to begin an IAM deployment? To find out the answers and to learn how ot build a comprehensive IAM roadmap, check out this presentation which discusses how Oracle can provide a quick start to your IAM program.  Platform approach-series-building a-roadmap-finalv1 View more presentations from OracleIDM

    Read the article

  • Ubuntu reports low battery capacity on my Dell Vostro

    - by Jeff
    I have a Dell Vostro 1500. Before I wiped Windows XP off my hard drive in 2009, I had a full ~7 hrs battery capacity. I installed Ubuntu 9, and the capacity immediately dropped to about 27% (and has since decreased to about 11%). I couldn't figure out what to do, so I've just lived with the 20-30 minute battery life ever since. I upgraded to Ubuntu 10, and the issue remained. I wiped my hard drive clean again and installed Ubuntu 11, and the issue still remains. I tried what they told me in the forum here, but it didn't do anything. Is it possible for a battery to suddenly lose most of its capacity?? Or is there a bug in the power management software?

    Read the article

  • Video Did Not Kill the Podcast Star

    - by Justin Kestelyn
    Who says video killed the podcast star? We're seeing more favorites out there than ever before. For example, the OTN team is proud to be supporters of the Java Spotlight Podcasts, straight from the official Java Evangelist Team at Oracle (lots of great insider info); the OurSQL: The MySQL Database Podcasts, produced by MySQL maven (and Oracle ACE Director) Sheeri Cabral; and The GlassFish Podcast, always a reliable source. And we'd add The Java Posse and The Basement Coders to our personal playlist. And although we're on a video kick ourselves at the moment, you can still get the audio of our TechCast Live shows, if you think we have "faces for radio."

    Read the article

  • ZFS for Database Log Files

    - by user12620111
    I've been troubled by drop outs in CPU usage in my application server, characterized by the CPUs suddenly going from close to 90% CPU busy to almost completely CPU idle for a few seconds. Here is an example of a drop out as shown by a snippet of vmstat data taken while the application server is under a heavy workload. # vmstat 1  kthr      memory            page            disk          faults      cpu  r b w   swap  free  re  mf pi po fr de sr s3 s4 s5 s6   in   sy   cs us sy id  1 0 0 130160176 116381952 0 16 0 0 0 0  0  0  0  0  0 207377 117715 203884 70 21 9  12 0 0 130160160 116381936 0 25 0 0 0 0 0  0  0  0  0 200413 117162 197250 70 20 9  11 0 0 130160176 116381920 0 16 0 0 0 0 0  0  1  0  0 203150 119365 200249 72 21 7  8 0 0 130160176 116377808 0 19 0 0 0 0  0  0  0  0  0 169826 96144 165194 56 17 27  0 0 0 130160176 116377800 0 16 0 0 0 0  0  0  0  0  1 10245 9376 9164 2  1 97  0 0 0 130160176 116377792 0 16 0 0 0 0  0  0  0  0  2 15742 12401 14784 4 1 95  0 0 0 130160176 116377776 2 16 0 0 0 0  0  0  1  0  0 19972 17703 19612 6 2 92  14 0 0 130160176 116377696 0 16 0 0 0 0 0  0  0  0  0 202794 116793 199807 71 21 8  9 0 0 130160160 116373584 0 30 0 0 0 0  0  0 18  0  0 203123 117857 198825 69 20 11 This behavior occurred consistently while the application server was processing synthetic transactions: HTTP requests from JMeter running on an external machine. I explored many theories trying to explain the drop outs, including: Unexpected JMeter behavior Network contention Java Garbage Collection Application Server thread pool problems Connection pool problems Database transaction processing Database I/O contention Graphing the CPU %idle led to a breakthrough: Several of the drop outs were 30 seconds apart. With that insight, I went digging through the data again and looking for other outliers that were 30 seconds apart. In the database server statistics, I found spikes in the iostat "asvc_t" (average response time of disk transactions, in milliseconds) for the disk drive that was being used for the database log files. Here is an example:                     extended device statistics     r/s    w/s   kr/s   kw/s wait actv wsvc_t asvc_t  %w  %b device     0.0 2053.6    0.0 8234.3  0.0  0.2    0.0    0.1   0  24 c3t60080E5...F4F6d0s0     0.0 2162.2    0.0 8652.8  0.0  0.3    0.0    0.1   0  28 c3t60080E5...F4F6d0s0     0.0 1102.5    0.0 10012.8  0.0  4.5    0.0    4.1   0  69 c3t60080E5...F4F6d0s0     0.0   74.0    0.0 7920.6  0.0 10.0    0.0  135.1   0 100 c3t60080E5...F4F6d0s0     0.0  568.7    0.0 6674.0  0.0  6.4    0.0   11.2   0  90 c3t60080E5...F4F6d0s0     0.0 1358.0    0.0 5456.0  0.0  0.6    0.0    0.4   0  55 c3t60080E5...F4F6d0s0     0.0 1314.3    0.0 5285.2  0.0  0.7    0.0    0.5   0  70 c3t60080E5...F4F6d0s0 Here is a little more information about my database configuration: The database and application server were running on two different SPARC servers. Storage for the database was on a storage array connected via 8 gigabit Fibre Channel Data storage and log file were on different physical disk drives Reliable low latency I/O is provided by battery backed NVRAM Highly available: Two Fibre Channel links accessed via MPxIO Two Mirrored cache controllers The log file physical disks were mirrored in the storage device Database log files on a ZFS Filesystem with cutting-edge technologies, such as copy-on-write and end-to-end checksumming Why would I be getting service time spikes in my high-end storage? First, I wanted to verify that the database log disk service time spikes aligned with the application server CPU drop outs, and they did: At first, I guessed that the disk service time spikes might be related to flushing the write through cache on the storage device, but I was unable to validate that theory. After searching the WWW for a while, I decided to try using a separate log device: # zpool add ZFS-db-41 log c3t60080E500017D55C000015C150A9F8A7d0 The ZFS log device is configured in a similar manner as described above: two physical disks mirrored in the storage array. This change to the database storage configuration eliminated the application server CPU drop outs: Here is the zpool configuration: # zpool status ZFS-db-41   pool: ZFS-db-41  state: ONLINE  scan: none requested config:         NAME                                     STATE         ZFS-db-41                                ONLINE           c3t60080E5...F4F6d0  ONLINE         logs           c3t60080E5...F8A7d0  ONLINE Now, the I/O spikes look like this:                     extended device statistics                  r/s    w/s   kr/s   kw/s wait actv wsvc_t asvc_t  %w  %b device     0.0 1053.5    0.0 4234.1  0.0  0.8    0.0    0.7   0  75 c3t60080E5...F8A7d0s0                     extended device statistics                  r/s    w/s   kr/s   kw/s wait actv wsvc_t asvc_t  %w  %b device     0.0 1131.8    0.0 4555.3  0.0  0.8    0.0    0.7   0  76 c3t60080E5...F8A7d0s0                     extended device statistics                  r/s    w/s   kr/s   kw/s wait actv wsvc_t asvc_t  %w  %b device     0.0 1167.6    0.0 4682.2  0.0  0.7    0.0    0.6   0  74 c3t60080E5...F8A7d0s0     0.0  162.2    0.0 19153.9  0.0  0.7    0.0    4.2   0  12 c3t60080E5...F4F6d0s0                     extended device statistics                  r/s    w/s   kr/s   kw/s wait actv wsvc_t asvc_t  %w  %b device     0.0 1247.2    0.0 4992.6  0.0  0.7    0.0    0.6   0  71 c3t60080E5...F8A7d0s0     0.0   41.0    0.0   70.0  0.0  0.1    0.0    1.6   0   2 c3t60080E5...F4F6d0s0                     extended device statistics                  r/s    w/s   kr/s   kw/s wait actv wsvc_t asvc_t  %w  %b device     0.0 1241.3    0.0 4989.3  0.0  0.8    0.0    0.6   0  75 c3t60080E5...F8A7d0s0                     extended device statistics                  r/s    w/s   kr/s   kw/s wait actv wsvc_t asvc_t  %w  %b device     0.0 1193.2    0.0 4772.9  0.0  0.7    0.0    0.6   0  71 c3t60080E5...F8A7d0s0 We can see the steady flow of 4k writes to the ZIL device from O_SYNC database log file writes. The spikes are from flushing the transaction group. Like almost all problems that I run into, once I thoroughly understand the problem, I find that other people have documented similar experiences. Thanks to all of you who have documented alternative approaches. Saved for another day: now that the problem is obvious, I should try "zfs:zfs_immediate_write_sz" as recommended in the ZFS Evil Tuning Guide. References: The ZFS Intent Log Solaris ZFS, Synchronous Writes and the ZIL Explained ZFS Evil Tuning Guide: Cache Flushes ZFS Evil Tuning Guide: Tuning ZFS for Database Performance

    Read the article

< Previous Page | 492 493 494 495 496 497 498 499 500 501 502 503  | Next Page >