Search Results

Search found 75452 results on 3019 pages for 'oracle big data'.

Page 860/3019 | < Previous Page | 856 857 858 859 860 861 862 863 864 865 866 867  | Next Page >

  • The good SQL database to process a lot of data?

    - by Dorian
    I have to process like 10-100 millions records. I have to give the data to the client when it's finish. The data is givent as SQL requests to execute in the database. He have a powerful server with MySQL, I think it will be fast enough. The issue is my computer is not as powerful as his server, so I would like to use an other SQL server who is compatible (I export his database and import it in my computer) with MySQL but more powerful. What should I use? Or am I doomed to use MySQL?

    Read the article

  • Multiple vulnerabilities in Firefox web browser

    - by chandan
    CVE DescriptionCVSSv2 Base ScoreComponentProduct and Resolution CVE-2012-0451 Improper Control of Generation of Code ('Code Injection') vulnerability 4.3 Firefox web browser Solaris 11 11/11 SRU 8.5 CVE-2012-0455 Improper Neutralization of Input During Web Page Generation ('Cross-site Scripting') vulnerability 4.3 CVE-2012-0456 Information Exposure vulnerability 5.0 CVE-2012-0457 Resource Management Errors vulnerability 9.3 CVE-2012-0458 Permissions, Privileges, and Access Controls vulnerability 6.8 CVE-2012-0459 Permissions, Privileges, and Access Controls vulnerability 7.5 CVE-2012-0460 Permissions, Privileges, and Access Controls vulnerability 6.4 CVE-2012-0461 Denial of Service (DoS) vulnerability 7.5 CVE-2012-0462 Denial of Service (DoS) vulnerability 7.5 CVE-2012-0464 Resource Management Errors vulnerability 7.5 This notification describes vulnerabilities fixed in third-party components that are included in Sun's product distribution.Information about vulnerabilities affecting Oracle Sun products can be found on Oracle Critical Patch Updates and Security Alerts page.

    Read the article

  • How can I write data to a file that users can't easily edit?

    - by ThePlan
    While working on game saving and loading I figured I could just write all the variable values to a file and then load that file from it's default location anytime. However from the very beginning it sounded like an odd job. I know about serialization and boost, but that seems so complicated, I figured I'd keep it simple, but I've ran across this huge issue: No matter what file I can write with C++, the user can get their hands on it, they can edit their position, they can remove a boss, or add new weapons for themselves. My question here is: How can I create a file in C++ which cannot be editted or openned with a text editor such as Notepad (I'm not trying to make a file which is impossible to open, but a file which will give the user a headache if he tries to edit it through usual methods.)

    Read the article

  • From Editor-in-Chief to Reader (part two)

    - by alexismp
    It's time for me (Alexis) to move to new challenges and this 773rd post of mine on TheAquarium will be my last one. For all these years since TheAquarium was started by others, this place has seen the community come to life, grow and thrive. It has been an effective way for the GlassFish team to share our progress and for you the community to provide feedback, both good and not so good. Thank you for your continued readership, support and feedback, all of which contributed to keep TheAquarium the #1 technical blog at Oracle and helped us improve Java EE and GlassFish in a significant way. Oracle has exciting and ambitious plans for GlassFish 4.0 and Java EE 7 and I trust TheAquarium will remain an important link in the chain as both of these are being delivered. I leave you in the hands of savvy and sharp contributors. Long live The Aquarium! - AlexisMP

    Read the article

  • Segment subdomains with Google Analytics?

    - by andrewpthorp
    So, when a website has multiple subdomains: www.example.com foo.example.com bar.example.com What is the best way to use Google Analytics to segment the data? I would prefer have access to 'All Data', 'Data from foo.example.com', and 'Data from bar.example.com'. I tried setting up 3 different views, and setting a filter on the foo/bar views that says: Include only traffic from the ISP domain that are equal to foo.example.com. However, I am not seeing any data collected into that View. I do, however, see all data in the 'All Data' view, but I can't figure out how to segment the data. I am including the analytics.js in the application.haml layout, which is always loaded in this app. Thanks!

    Read the article

  • What Can I Use A Laptop Data Card For?

    So you?ve been looking at alternatives to having internet cable installed in your home and you?re feeling pretty guilty about stealing the neighbors Wi-Fi. If you?re tired of going down to your local... [Author: Tony Previte - Computers and Internet - April 03, 2010]

    Read the article

  • New Solaris Cluster!

    - by Jeff Victor
    We released Oracle Solaris Cluster 4.1 recently. OSC offers both High Availability (HA) and also Scalable Services capabilities. HA delivers automatic restart of software on the same cluster node and/or automatic failover from a failed node to a working cluster node. Software and support is available for both x86 and SPARC systems. The Scalable Services features manage multiple cluster nodes all providing a load-balanced service such as web servers or app serves. OSC 4.1 includes the ability to recover services from software failures, failure of hardware components such as DIMMs, CPUs, and I/O cards, a global file system, rolling upgrades, and much more. Oracle Availability Engineering posted a brief description and links to details. Or, you can just download it now!

    Read the article

  • Using all Ten IO slots on a 7420

    - by user12620172
    So I had the opportunity recently to actually use up all ten slots in a clustered 7420 system. This actually uses 20 slots, or 22 if you count the clusteron card. I thought it was interesting enough to share here. This is at one of my clients here in southern California. You can see the picture below. We have four SAS HBAs instead of the usual two. This is becuase we wanted to split up the back-end taffic for different workloads. We have a set of disk trays coming from two SAS cards for nothing but Exadata backups. Then, we have a different set of disk trays coming off of the other two SAS cards for non-Exadata workloads, such as regular user file storage. We have 2 Infiniband cards which allow us to do a full mesh directly into the back of the nearby, production Exadata, specifically for fast backups and restores over IB. You can see a 3rd IB card here, which is going to be connected to a non-production Exadata for slower backups and restores from it.The 10Gig card is for client connectivity, allowing other, non-Exadata Oracle databases to make use of the many snapshots and clones that can now be created using the RMAN copies from the original production database coming off the Exadata. This allows for a good number of test and development Oracle databases to use these clones without effecting performance of the Exadata at all.We also have a couple FC HBAs, both for NDMP backups to an Oracle/StorageTek tape library and also for FC clients to come in and use some storage on the 7420.  Now, if you are adding more cards to your 7420, be aware of which cards you can place in which slots. See the bottom graphic just below the photo.  Note that the slots are numbered 0-4 for the first 5 cards, then the "C" slots which is the dedicated Cluster card (called the Clustron), and then another 5 slots numbered 5-9. Some rules for the slots: Slots 1 & 8 are automatically populated with the two default SAS cards. The only other slots you can add SAS cards to are 2 & 7. Slots 0 and 9 can only hold FC cards. Nothing else. So if you have four SAS cards, you are now down to only four more slots for your 10Gig and IB cards. Be sure not to waste one of these slots on a FC card, which can go into 0 or 9, instead.  If at all possible, slots should be populated in this order: 9, 0, 7, 2, 6, 3, 5, 4

    Read the article

  • Multiple vulnerabilities in Thunderbird

    - by chandan
    CVE DescriptionCVSSv2 Base ScoreComponentProduct and Resolution CVE-2012-0451 Improper Control of Generation of Code ('Code Injection') vulnerability 4.3 Thunderbird Solaris 11 11/11 SRU 8.5 CVE-2012-0455 Improper Neutralization of Input During Web Page Generation ('Cross-site Scripting') vulnerability 4.3 CVE-2012-0456 Information Exposure vulnerability 5.0 CVE-2012-0457 Resource Management Errors vulnerability 9.3 CVE-2012-0458 Permissions, Privileges, and Access Controls vulnerability 6.8 CVE-2012-0459 Permissions, Privileges, and Access Controls vulnerability 7.5 CVE-2012-0460 Permissions, Privileges, and Access Controls vulnerability 6.4 CVE-2012-0461 Denial of Service (DoS) vulnerability 7.5 CVE-2012-0462 Denial of Service (DoS) vulnerability 7.5 CVE-2012-0464 Resource Management Errors vulnerability 7.5 This notification describes vulnerabilities fixed in third-party components that are included in Sun's product distribution.Information about vulnerabilities affecting Oracle Sun products can be found on Oracle Critical Patch Updates and Security Alerts page.

    Read the article

  • Is it reasonable to make a RAID-1 array with a ram disk and a physical disk to maximize read performance and protect data?

    - by Petr Pudlák
    In one of the answers on SO (I forgot which one) I've seen a suggestion to make a RAID-1 array composed of a RAM disk and a physical partition. By adding the physical partition with --write-mostly and enabling --write-behind the system should read everything instantly from the RAM disk but still save all data to the physical partition so that the data are preserved and the RAID array can be assembled again after reboot. Is such a setup reasonable? Will it perform any better in some scenario than having just the physical partition and perhaps tweaking the kernel to favor disk cache (swappiness and vfs_cache_pressure)?

    Read the article

  • Changing Word mail merge data source locations in bulk?

    - by Daft Viking
    I've just moved a number of Word mail merge files, and a number of Excel spreadsheets that are the data sources for the mail merges, from a Windows XP computer to a Windows 7 computer, and now all the paths for the merge sources are incorrect (used to be c:\documents and settings\user\my documents.... now c:\users\documents....). While I can correct the path of the data source in each file individually, I was hoping that there would be some way of updating the files in bulk, as there are a relatively large number of them. Word 2007 is what is being used, but the documents are all in the previous DOC format (not DOCX).

    Read the article

  • How do I add a counter for process data in perfmon.exe for a process that isn't currently running?

    - by Jason Jones
    I would like to perform an ad hoc capture of data using perfmon.exe for a process that runs during the night. I know the name of the executable--lets call it Foo.exe. If the process were currently running, I would go to the Add Counters dialog in perfmon, switch to the Process object, and select the Foo instance from the list. However, it's not currently running. Is it possible to set up perfmon so that it will capture process data for this process when it starts, and if so, how would I configure it to do so?

    Read the article

  • Not Just What You Sell but Really How You Sell?

    - by divya.malik
    Sales 2.0 is changing the way customers get influenced and buy the products that they want to purchase. Engaging and listening to your customers and market  regularly is necessary, since they are engaging with other customers like themselves all the time, and online. Creating a consistent customer experience for your customers, across channels is more critical than ever. 2.0 and Social media as a channel need to become a part of your inbound and outbound sales strategy. Oracle has been investing in new capabilities to address the needs of this changing marketplace. Listen to  Mark Woollen, VP of CRM at Oracle discuss these new innovations that are changing the way companies and customer interact today, and the new strategies that will give you the lead in the marketplace. Here is an excerpt from his presentation (, which was featured on SellingPower.com.

    Read the article

  • Trying to grok Linux quotas, where is the data stored?

    - by CarpeNoctem
    So all the tutorials and documentation for the Linux quota system has left me confused. For each filesystem with quotas enabled/on where is the actual quota information stored? Is it filesystem metadata or is it in a file? Say user foo creates a new file on /home. How does the kernel determine whether user foo is below their hard limit? Does the kernel have to tally up quota information on that filesystem each time or is it in the superblock or somewhere else? As far as I understand, the kernel consults the aquota.user file for the actual rules, but where is the current quota usage data stored? Can this be viewed with any tools outside repquota and the like? TIA!! Update: Thanks for the help. I had already read that mini-HOWTO. I am pretty clear on the usage of the user space tools. What I was unclear on is whether the usage data was ALSO in the file that stored per-user limits and you answered this with a yes. From what I can tell, rc.sysinit runs quotacheck and quotaon on startup. The quotacheck program analyzes the filesystem, updates the aquota.* files. It then makes use of quota.h and the quotactl() syscall to inform the kernel of quota info. From this point forward the kernel hashes that information and increments/decrements quota stats as changes occur. Upon shutdown, the init.d/halt script runs the quotaoff command RIGHT before the filesystems are unmounted. The quotaoff command does not appear to update the aquota.* files with the information the kernel has in memory. I say this because the {a,c,m}times for the aquota.user file are only updated upon a reboot of the system or by manual running the quotacheck command. It appears - as far as I can tell - that the kernel just drops it's up-to-date usage data on the floor at shutdown. This information is never used to update the aquota.* files. They are updated during startup by quotacheck(rc.sysinit). Seems silly to me since that updated info had already been collected by the kernel. So...in conclusion I am still not entirely clear on the methods. ;)

    Read the article

  • What's the correct approach for passing data from several models into a service?

    - by Doug Chamberlain
    I have an AccountModel and a page where the user can upload a file. What I would like to have happen is when the user uploads the file. The PageController does something like the following. this is a quick attempt just written in the question to illustrate my question. public class PageController : Controller { private Service service; public ActionResult Upload(HttpPostedFileBase f){ service.savefile(f,_AccountModel_whatever.currentlyloggedinuser.taxid) } } public class Service { // abunch of validation and error checking to make sure the file is good to store } Wouldn't this approach be in bad practice? Since I'm making my controller dependent on the existence of th AccountModel? This will become a HUGE program over the next few years, and I really want to maximize the quality of the framework now.

    Read the article

  • Java EE@Java Day Taipei 2014

    - by reza_rahman
    Java Day Taipei 2014 was held at the Taipei International Convention Center on August 1st. Organized by Oracle University, it is one of the largest Java developer events in Taiwan. This was another successful year for Java Day Taipei with a fully sold out venue. In addition to Oracle speakers like me, Steve Chin and Naveen Asrani, the event also featured a bevy of local speakers including Taipei Java community leaders. Topics included Java SE, Java EE, JavaFX and Big Data. I delivered a keynote on Java EE 7/Java EE 8 as well as a talks on aligning the JavaScript ecosystem with Java EE 7 and using NoSQL solutions in Java EE applications. More details on the sessions and Java Days Taipei, including the slide decks and code, posted on my personal blog.

    Read the article

  • Engineered Systems Announcements at OOW

    - by Cinzia Mascanzoni
    Normal 0 false false false EN-US X-NONE X-NONE MicrosoftInternetExplorer4 /* Style Definitions */ table.MsoNormalTable {mso-style-name:"Table Normal"; mso-tstyle-rowband-size:0; mso-tstyle-colband-size:0; mso-style-noshow:yes; mso-style-priority:99; mso-style-qformat:yes; mso-style-parent:""; mso-padding-alt:0cm 5.4pt 0cm 5.4pt; mso-para-margin:0cm; mso-para-margin-bottom:.0001pt; mso-pagination:widow-orphan; font-size:10.0pt; font-family:"Calibri","sans-serif"; mso-bidi-font-family:"Times New Roman";} See below the  announcements of new Exa systems at OOW Oracle Exadata X3 Database In-Memory Machine and  Oracle Exalogic Elastic Cloud X3-2 announcements.

    Read the article

  • How do I scale EC2 and push out code / data to my instances?

    - by chris
    Unfortunately I only have a limited knowledge of server architecture, I come from a development background. I am looking to ensure my new app can scale properly using EC2. I currently have a T1.micro for development running Windows with SQL server 2008. The system allows students to come to our site to search for a mentor, update their profile with pictures and employment history etc. Roughly the same sort of work as a LinkedIn profile. I need this to be able to scale very quickly without wasted resources. I understand the following is important. Separation of data, application etc. I will achieve this I think by hosting images using S3, Database instance via RDS and upgrade the EC2 instance. My main question is: How do I push data / code out to multiple ec2 / RDS instances seamlessly?

    Read the article

  • The Role of High Availability Computing on Business Continuity -- Part 2 of 2

    For organizations that can't afford, sustain or justify downtime -- developing, implementing and testing a high-availability computing strategy is essential. Unplanned downtime affects company reputation, stock price and competitive strategy. It can even delay IT innovation projects necessary for delivering new services to customers. Learn how Oracle's approach to high availability computing is fundamentally different from the traditional model. Hear Oracle Thought Leader Balaji Bashyam (Vice President, Global Database Support) discuss high availability strategy, best practices, and the effects of availability on business, in a question and answer interview format. This podcast is presented in two parts and is intended for an audience of decision makers and influencers.

    Read the article

  • Leveraging Existing ERP Systems to Support Environmental Accounting and Reporting

    Organizations globally are faced with a complex set of emissions reporting requirements. Driven by country-specific regulatory mandates as well as stakeholder requests for voluntary reporting, companies are under pressure to provide consistent, transparent and accurate collection, measurement and reporting of energy usage and emissions data. In this podcast, you'll year about how the new Oracle Environmental Accounting and Reporting solution extends the capabilities of Oracle E-Business Suite and JD Edwards Financials to enable organizations to track their greenhouse gas emissions and other environmental data against reduction targets, and to obtain accurate, repeatable and verifiable methodologies for greenhouse gas calculation in accordance with global standards and for both voluntary and legislated emissions reporting schemes.

    Read the article

  • So funktioniert Marketing!

    - by A&C Redaktion
    Ein Großteil des Oracle Business wird von Partnern getragen. Damit das erfolgreich funktioniert, unterstützt Alliances & Channel die Partner mit gezielten Marketing-Aktionen. Wie zum Beispiel der Oracle Exadata Kampagne. Mit ausgewählten, spezialisierten Exadata-Partnern wurde die europaweit durchgeführten Demand Generation-Kampagne umgesetzt: Neben einem Emailing, das die jeweilige Exadata-Lösung des Partners im Fokus hatte, wurde für jeden Partner eine adaptierte Microsite erstellt und im Anschluss von der Marketingagentur nachtelefoniert. Die daraus entstandenen Leads potenzieller Kunden wurden dann an die Partner zurückgegeben, um den Vertriebsprozess zu starten. Die beiden deutschen Teilnehmer der Kampagne, inforsacom und ISE, berichten in der aktuellen Ausgabe von Partner Insight ausführlich über Ihre Erfahrungen und Erfolge der Exadata Kampagne.

    Read the article

< Previous Page | 856 857 858 859 860 861 862 863 864 865 866 867  | Next Page >