Search Results

Search found 51747 results on 2070 pages for 'oracle database in memory'.

Page 463/2070 | < Previous Page | 459 460 461 462 463 464 465 466 467 468 469 470  | Next Page >

  • Passenger/Rails not releasing memory

    - by michaeldelorenzo
    I have an Ubuntu server running three separate Rails (2.3.8) applications with Passenger, REE and Apache. Recently we started experiencing problems with ruby processes eating up memory and consuming entire cores on our server. Here's what we're getting... %CPU PID USER COMMAND 99.9 1717 nobody Rails: /var/www/api 99.6 5542 nobody Rails: /var/www/api 97.3 1223 nobody Rails: /var/www/api 4.7 5537 nobody Passenger ApplicationSpawner: /var/www/api 10.5 1801 nobody Rails: /var/www/api We've also seen instances where there have been over 100 instances of Apache running. These applications have been running for a few months without an of these issues, but in the last day or so we've been noticing this. The site referenced here is a Rails application that is a RESTful API so it serves many requests every minute. Any guidance on what we should be checking or looking out for would be appreciated.

    Read the article

  • Extra, Extra, Read All About It- Offer Ends Soon!

    - by Kristin Rose
    Normal 0 false false false EN-US X-NONE X-NONE /* Style Definitions */ table.MsoNormalTable {mso-style-name:"Table Normal"; mso-tstyle-rowband-size:0; mso-tstyle-colband-size:0; mso-style-noshow:yes; mso-style-priority:99; mso-style-qformat:yes; mso-style-parent:""; mso-padding-alt:0in 5.4pt 0in 5.4pt; mso-para-margin-top:0in; mso-para-margin-right:0in; mso-para-margin-bottom:10.0pt; mso-para-margin-left:0in; line-height:115%; mso-pagination:widow-orphan; font-size:11.0pt; font-family:"Calibri","sans-serif"; mso-ascii-font-family:Calibri; mso-ascii-theme-font:minor-latin; mso-fareast-font-family:"Times New Roman"; mso-fareast-theme-font:minor-fareast; mso-hansi-font-family:Calibri; mso-hansi-theme-font:minor-latin; mso-bidi-font-family:"Times New Roman"; mso-bidi-theme-font:minor-bidi;} Start spreading the news... Your partner news of course by submitting all interesting presentation ideas to the Oracle OpenWorld 2012 Call for Papers. Though you may not be able to serenade your customers with a voice like Sinatra’s, you can still get their attention by sharing your customer solutions, highlighting your achievements and attempting your best “Old Blue Eyes” impersonation. This call for papers will end April 9th, 2012 so don’t be a stranger in the night; instead fly your company to the moon and back by getting those papers in. May luck be a lady or simply on your side, as all accepted submission speakers will receive a complimentary pass to the event they have been accepted for. Yes you’re lovely, so why wait any longer? Join the Oracle OpenWorld 2012 ‘Rat Pack’ today by watching the video below or submitting to the call for papers. The best is yet to come, The OPN Communications Team

    Read the article

  • How much faster is a database running in RAM?

    - by orokusaki
    I"m looking to run PostgreSQL in RAM for performance enhancement. The database isn't more than 1GB and shouldn't ever grow to more than 5GB. Is it worth doing? Are there any benchmarks out there? Is it buggy? My second major concern is: How easy is it to back things up when it's running purely in RAM. Is this just like using RAM as tier 1 HD, or is it much more complicated?

    Read the article

  • Which would be better? Storing/access data in a local text file, or in a database?

    - by TerranRich
    Basically, I'm still working on a puzzle-related website (micro-site really), and I'm making a tool that lets you input a word pattern (e.g. "r??n") and get all the matching words (in this case: rain, rein, ruin, etc.). Should I store the words in local text files (such as words5.txt, which would have a return-delimited list of 5-letter words), or in a database (such as the table Words5, which would again store 5-letter words)? I'm looking at the problem in terms of data retrieval speeds and CPU server load. I could definitely try it both ways and record the times taken for several runs with both methods, but I'd rather hear it from people who might have had experience with this. Which method is generally better overall?

    Read the article

  • database independent coding framework options?

    - by statirasystems
    Background: I have not programmed in a while besides doing VBA and a little VB.NET. So please forgive my language use. I'm green and have a head cold. I am reading all I can now, but I have no programming circles to draw from. The information I am providing is to help guide you to what I am looking for. I am not confident I can ask the question properly. Story: I have four different projects that I am starting. Obviously I won't be working on all at the same time however they each will have similar needs and be inter related. They are as follows: Desktop Environment/System User Interface - basically a product that runs on major computers via mono or .net that unifies the look and functions. In the context of the up coming question it would be able to directly access data of various types. It would work in tandum with my office suite, system manager, and network application framework. Office Suite - technically it would not be a suite since I will be doing it from one interfacel except for the Communications Application. As far as the question, it will need to be able to link to various data sources for storing files and using, manipulating, and presenting information. System Manager - an intellegent system to manage and administer the entire network and all equipment. As far as the question, needs to be able to access data for archiving and and for accessing it's own settings stored in various formats, sql or xml. Network Application Framework - A complete system that can be used for ERP, CRM, CMS, Errata, File Management, and so on. As to the question to be able to access it's own or interlink with existing applications. Requirement: C#, Simplifies and reduces coding, use the same code to access diffent databases(ie MySQL, MS SQL, ACCESS, XML, ...), Mono would be nice but not a must, Question: What librarys, frameworks, or other options would be able to help with this? Is there a good resource to guide me? I don't want arguing over what is best, just information to help me further understand and make an educated decision.

    Read the article

  • Android googlemap Out of memory

    - by Xiaofeng
    Hi, I made an android application with googlemap api, and draw some 16x16 png (about 200 count) on overlay. When I move or zoom on/in mapview, "out of memory" error occurs very often. I also used the googlemap appication in my htc itself. Seams that it uses about 14+MB memmory, and never happens "out of memory". How to save memmory usage in a googlemap api, or how to enlarge android api memmory limit. Thanks a lot!

    Read the article

  • Clearing Windows file share "memory"

    - by Tom Shaw
    I'm currently upgrading a Samba file server (from 3.0.23d to 3.4.3). I have a problem on the Windows client side: if the client was accessing a UNC path or mapped drive from the Samba server before the upgrade, then after the upgrade those paths or drives are not accessible. However, I can consistently resolve the client side problem for good by rebooting the client and then re-mapping all of the mapped drives. The problem appears to be related to the client's "memory" of the pre-upgrade Samba server, which the reboot and re-map clears. I have the same issue and same fix on Windows XP SP3 and Windows Server 2003 SP2. This question is specifically: is it possible to reproduce the benefits of the Windows reboot without actually rebooting the client? I have tried restarting various Windows services, disabling and enabling the network, logging out and back in again, but nothing except a reboot appears to do the trick.

    Read the article

  • Can I get memory usage of processes running on the monitored server by newrelic REST API

    - by youlin
    according to the newrelic faq https://docs.newrelic.com/docs/server/server-monitor-faq, The Server Monitoring agent can report Top 20 processes that are using significant memory or I/O and I can view the memory usage of the processes on the newrelic portal page. However, I do not find any clue about how to get this metrics by newrelic REST API (I can get the CPU usage of processes by REST API). Is it possible to do this?

    Read the article

  • get-wmiobject sql join in powershell - trying to find physical memory vs. virtual memory of remote s

    - by Willy
    get-wmiobject -query "Select TotalPhysicalMemory from Win32_LogicalMemoryConfiguration" -computer COMPUTERNAME output.csv get-wmiobject -query "Select TotalPageFileSpace from Win32_LogicalMemoryConfiguration" -computer COMPUTERNAME output.csv I am trying to complete this script with an output as such: Computer Physical Memory Virtual Memory server1 4096mb 8000mb server2 2048mb 4000mb

    Read the article

  • Memory leaking issue iPhone

    - by milanjansari
    Hello, Anyone have a idea about solve memory leak issues i have found one memory related issue NSAutoreleaseNoPool(): Object 0x3588aea0 of class NSCFString autoreleased with no pool in place - just leaking can anyone have a idea about how can i solve.. Thnak you

    Read the article

  • Would you store binary data in database or in file system?

    - by paul
    This is a question which has been asked before (large-text-and-images-in-sql) but mainly for data which will be changed. In my case the data will be stored and never changed. Just seems sensible to keep everything together. Are there any reasons why I should not store static binary data in a database? Assuming it is a sensible thing to do, are there any advantages to storing such data in separate tables? (You might begin to realise now that I'm not a DB expert...) Clarify: There will probably be no more than 10-20 users but these will be in the US and in the UK. The binary data will have to be transfered in any case.

    Read the article

  • Obtaining memory available to JVM at runtime

    - by Bo Tian
    I'm trying to sort a bunch of data such that that the size of data input to the program can be larger than the memory available to the JVM, and handling that requires external sort which is much slower than Quicksort. Is there any way of obtaining memory available to the JVM at runtime such that I could use in place sorting as much as possible, and only switch to Mergesort when data input is too large?

    Read the article

  • What amount of PHP memory usage is reasonable for multiple WordPress sites? [closed]

    - by John
    It seems like I've been plagued with memory usage issues most of the time I've been blogging (7 years). I upgraded my hosting to a HostGator VPS and still seem to have issues. I have about a dozen WP 3.4.2 sites running within a single account. Each has a separate installation. I finally installed TPC! Memory Usage within one of my WP installs to see what was going on. The memory usage just standing still is on the order of 40 MB. This seems high, but I don't know. What would be a reasonable memory usage for these WP installs? Do I need to take other things into account, or am I even looking at it the right way?

    Read the article

  • How does Google store search trends in backend?

    - by Achshar
    Google trends shows what query has been searched how many times and some other properties of the said query. But how is this data stored in a database? Storing a new row for every search does not seem right. They also tell the query on a time graph, so they must have some way to look for individual searches made by users, but the number of queries they get every day, it does not feel right that they would store every search in a database row along with a time-stamp. This does not apply to just Google trends or Google in general but any other big site that gets awful number of queries and then has tools to see them in depth. I am not an expert on this but I am interested to know some high level structure of how things work behind the scenes.

    Read the article

  • sqlite error :/* SQL error or missing database */

    - by user262325
    Hello everyone I have a project in which I stored sqlite database file "data.sqlite3" to 'Group'&files'-'resource' Below are my viewcontroller source codes //-myviewcontroller.h #import "sqlite3.h" #define kFilename @"data.sqlite3" //myviewcontroller.m -(NSString *)dataFilePath { NSArray *paths = NSSearchPathForDirectoriesInDomains( NSDocumentDirectory, NSUserDomainMask, YES); NSString *documentsDirectory = [paths objectAtIndex:0]; return [documentsDirectory stringByAppendingPathComponent:kFilename]; } -(void)f { if (sqlite3_open([[self dataFilePath] UTF8String],&database)!=SQLITE_OK) //dataFilePath returns ///Users/interdev/Library/Application Support/iPhone Simulator/User/Applications/095C6E05-4EAE-4817-883E-A72E39D439E0/Documents/data.sqlite3 { sqlite3_close(database); NSAssert(0,@"Failed to open database");//no problem } NSString *query = @"SELECT * FROM table1 ORDER BY ROW";//table1 is table name sqlite3_stmt *statement; NSInteger v=sqlite3_prepare_v2( database, [query UTF8String], -1, &statement, nil); NSString *zs= [NSString stringWithFormat:@"%d",v]; NSLog(@" The buttontitile is %@ ",zs); if ( v == SQLITE_OK) { // ... } I checked value of v in log, it always is 1 #define SQLITE_ERROR 1 /* SQL error or missing database */ I do not know why this happened.

    Read the article

  • Deal Registration is Moving to the Oracle Partner Store!

    - by Cinzia Mascanzoni
    In November 2012, Oracle will unveil a new partner deal registration system within Oracle Partner Store (OPS). At that time, OPS will become the single source for partners to register deals, obtain deal status, and place orders. The new deal registration system will offer several enhancements, including: Simplified Registration Form Easier Product Selection Expanded Browser Support Shared Registration Visibility Between VAD and VAR Pre-set Customer Selection from Partner Ordering Base Read more here.

    Read the article

  • Allocation of memory with LE routines

    - by HaWe
    At /questions/2000777/allocation-of-memory-in-variable-length-tables NealB mentioned LE routines to allocate/deallocate memory in a non-CICS COBOL program. I'd very much like to know how this is done: how the LE routine is called. (I'm familiar with the LINKAGE SECTION and with SET ADDRESS.) Since I have no access to an IBM mainframe at the moment - meaning no access to online documentation - some code snippets could enlighten me.

    Read the article

  • How to cast a pointer of memory block to std stream

    - by Shahrooz Kia
    I have programed an application on windows XP and in Visual Studio with c++ language. In that app I used LoadResource() API to load a resource for giving a file in the resource memory. It returned a pointer of memory block and I wanna cast the pointer to the std stream to use for compatibility. Could anyone help me?

    Read the article

  • How can I efficiently retrieve a large number of database settings as PHP variables?

    - by Steven
    Currently all of my script's settings are located in a PHP file which I 'include'. I'm in the process of moving these settings (about 100) to a database table called 'settings'. However I'm struggling to find an efficient way of retrieving all of them into the file. The settings table has 3 columns: ID (autoincrements) name value Two example rows might be: 1 admin_user john 2 admin_email_address [email protected] The only way I can think of retrieving each setting is like this: $result = mysql_query("SELECT value FROM settings WHERE name = 'admin_user'"); $row = mysql_fetch_array($result); $admin_user = $row['value']; $result = mysql_query("SELECT value FROM settings WHERE name = 'admin_email_address'"); $row = mysql_fetch_array($result); $admin_email_address = $row['value']; etc etc Doing it this way will take up a lot of code and will likely be slow. Is there a better way? Thanks.

    Read the article

  • What is the best way to keep database data encrypted with user passwords?

    - by Dan Sosedoff
    Let's say an application has really specific data which belongs to a user, and nobody is supposed to see it except the owner. I use MySQL database with DataMapper ORM mapper. The application is written in Ruby on Sinatra. Application behavior: User signs up for an account. Creates username and password. Logs into his dashboard. Some fields in specific tables must be protected. Basically, I'm looking for auto-encryption for a model properties. Something like this: class Transaction include DataMapper::Resource property :id, Serial property :value, String, :length => 1024, :encrypted => true ... etc ... belongs_to :user end I assume that encryption/decryption on the fly will cause performance problems, but that's ok. At least if that works - I'm fine. Any ideas how to do this?

    Read the article

  • How to migrate Django models from mysql to sqlite (or between any two database systems)?

    - by Daphna Shezaf
    I have a Django deployment in production that uses MySQL. I would like to do further development with SQLite, so I would like to import my existing data to an SQLite database. I There is a shell script here to convert a general MySQL dump to SQLite, but it didn't work for me (apparently the general problem isn't easy). I figured doing this using the Django models must be much easier. How would you do this? Does anyone have any script to do this?

    Read the article

< Previous Page | 459 460 461 462 463 464 465 466 467 468 469 470  | Next Page >