Search Results

Search found 18811 results on 753 pages for 'dynamic memory allocation'.

Page 423/753 | < Previous Page | 419 420 421 422 423 424 425 426 427 428 429 430  | Next Page >

  • What are the differences between the "generic" and "server" kernel images provided by Ubuntu?

    - by dcrosta
    In particular, I'm wondering if there are any patches or config adjustments made to the disk cache size in the server edition. I'm running on a small system (256M RAM), and would like to experiment with keeping the disk cache size smaller so that there's more memory available for applications. I've found this page at Ubuntu's website, which neither answers my questions nor is about the 9.04 release.

    Read the article

  • Low FPS in some games, but hardware not fully used

    - by Mario De Schaepmeester
    I just did a little funny experiment in the game/sim "Train Simulator 2013". I normally have good FPS in it (around 30) at full settings. What I did was make a really, really long train so that the calculations the sim needed to make were enormous (the sim is quite realistic, it takes all things into account like speed/acceleration, G-forces, comfort levels, possible wheel slip and many more, and most of those things on each carriage seperately). This resulted in only 14FPS as reported by the game, but it felt more like 8FPS or so. I have a Logitech G15 keyboard which has an LCD, and it allows me to monitor CPU/RAM and video card load on it. The strange thing is, all CPU cores were busy, but the total load was only about 60% maximum at all times. The video card was only on 30% load (possibly an important note, the memory was full, which is however not unusual for the game in question). The RAM had plenty of room and there weren't many operations as it didn't grow or shrink much. I just have the feeling that the game would run smoother if it used more of my hardware power. Why is it not doing so? I had the same in another game, The Elder Scrolls: Morrowind when using more than 100 mods (that all use scripting) and a few high res texture mods, + a full-on graphics improvement program. The engine is very old (2003), and so I thought this might be the cause (not being optimised for multithreading). I had thought of possible causes, like: The operating system doesn't let the games use all the resources. It doesn't make use of multi-threading appropriately. To eliminate the former, I tried a CPU stress tool and that got 100% CPU juice as I let it run, so the OS is not the problem. I gave its thread the "higher" priority though. My actual question In both games, I did things the engine was not really built to do or support. Can those games' framerate be limited cause of their own engine not being able to cope? What is the real reason and more importantly, can I help it? And in any case, could something actually be wrong with my hardware? It's all reasonably new, a couple of months, and I (almost) never experience any other trouble. Modern and much more demanding games work absolutely fine. Specs CPU: AMD Phenom II 965 X4 @ 3.4gHz RAM: 8GB of DDR3 RAM Video: MSI GTX560 (nVidia chip) with 1GB of GDDR5 memory OS: Windows 7 Ultimate 64 bit Nothing overclocked.

    Read the article

  • How to find the process(es) which are hogging the machine

    - by Aaron Digulla
    Scenario: All of a sudden, my computer feels sluggish. Mouse moves but windows take ages to open, etc. uptime says the load is 7.69 and raising. What is the fastest way to find out which process(es) are the cause of the load? Now, "top" and similar tools isn't the answer because they either show CPU or memory usage but not both at the same time. What I need is the single command which I might be able to type as it happens - something that will figure out any of System is trying to swap 8GB of RAM to disk because process X ... or process X seeks all over the disk or process X uses 400% CPU" So what I'm looking for is iostat, htop/atop and similar tools run into one with an output like this: 1235 cp - Disk trashing 87 chrome - Uses 2&nbsp;GB of RAM 137 nfs_bench - Uses 95% of the network bandwidth I don't want a tool that gives me some numbers which I can analyze but a tool that tells me exactly which process causes the current load. Assume that the user in front of the keyboard barely knows how to write "process", but the user is quickly overwhelmed when it comes to "resident size", "virtual memory" or "process life cycle". My argument goes like this: A user notices a problem. There can be thousands of reasons ... well, almost :-) The user wants to know the source of the problem. The current solutions give me lots of numbers, and I need to know what these numbers mean. What I'm looking for is a meta tool. 99% of the data is irrelevant to the problem. So what the tool should do is look for processes which hog some resource and list only those along with "this process needs a lot of CPU, this produces many IRQs, this process allocates a lot of RAM (and it's still growing)". This will be a relatively short list. It will be much more simple for someone new to this to locate the culprit from this list than from the output of, say, htop which gives me about 5000 numbers but requires me to fold multi-threaded processes myself (I have 50 lines which say VIRT 2750M but only 16 GB of RAM - the machine ought to swap itself to death but of course, this is a misinterpretation of the data that can happen quickly).

    Read the article

  • A Versatile Physical Server

    - by Paul
    How does one judge potential memory and processor needs for linux web servers? Specifically, given: A Debian or Ubuntu OS Running a web server (apache2), and A database (MySQL), and A DNS server (bind), and Being used by up to 100 concurrent users, at some points each downloading high-resolution (0.5 to 1 MB) images via a web app. How much should one budget in terms of RAM, type of processor(s), and number of cores? Thanks!

    Read the article

  • Almost All Logical Volumes Disappeared - Recovery?

    - by Alex
    We had a hard disc crash of one of two hard discs in a software raid with a LVM on top. The server is running Citrix xenserver. On the hard disk which is still intact, the volume group gets detected well, but only one LV is left. (some hashes replaced by "x") # lvdisplay --- Logical volume --- LV Name /dev/VG_XenStorage-x-x-x-x-408b91acdcae/MGT VG Name VG_XenStorage-x-x-x-x-408b91acdcae LV UUID x-x-x-x-x-x-vQmZ6C LV Write Access read/write LV Status available # open 0 LV Size 4.00 MiB Current LE 1 Segments 1 Allocation inherit Read ahead sectors auto - currently set to 256 Block device 253:0 root@rescue ~ # vgdisplay --- Volume group --- VG Name VG_XenStorage-x-x-x-x-408b91acdcae System ID Format lvm2 Metadata Areas 1 Metadata Sequence No 4 VG Access read/write VG Status resizable MAX LV 0 Cur LV 1 Open LV 0 Max PV 0 Cur PV 1 Act PV 1 VG Size 698.62 GiB PE Size 4.00 MiB Total PE 178848 Alloc PE / Size 1 / 4.00 MiB Free PE / Size 178847 / 698.62 GiB VG UUID x-x-x-x-x-x-53w0kL I could understand if a full physical volume is lost - but why only the logical volumes? Is there any explanation for this? Is there any way to recover the logical volumes? EDIT We are here in a rescue system. The problem is that the whole server does not boot (GRUB error 22) What we are trying to do is to access the root filesystem. But everything was in the LVM. We have only this: (parted) print Model: ATA SAMSUNG HD753LJ (scsi) Disk /dev/sdb: 750GB Sector size (logical/physical): 512B/512B Partition Table: msdos Number Start End Size Type File system Flags 1 32.3kB 750GB 750GB primary boot, lvm And this 750GB LVM volume is exactly what we see on top.

    Read the article

  • 256MB VPS on XEN

    - by user63410
    I am thinking of getting a 256MB server on XEN but not sure is it capable of handling this setup: Varnish + Nginx + Eaccelerator + PHP-FM + MySQL + Mail/FTP? I've tried this setup on OpenVZ and it was partly disastorous, especially with varnish in the equation, i've heard that XEN is better at memory management, sometimes shaving off 50mb or more in comparison to OpenVZ setups... if anyone has any helpful suggestions / input please let me know thanks

    Read the article

  • iSCSI NAS supporting VMware Esxi 5.1 servers in a software development environment

    - by BigTFromAZ
    I would like to build (or purchase) the smallest, quietest most energy efficient servers that I can for a low demand environment. These would only have a small disk for guest swap files, a CPU (XEON preferred), 32 Gb of memory and minimal graphics. The virtual machines need to be up but demand will be quite low. Any thought on case, format processors, fans, motherboard, et al? Small, quiet and lots of RAM are the operative words here.

    Read the article

  • Need help in using svn on ubuntu 9.10

    - by michael
    Hi, I have install svn on ubuntu 9.10. But when I try to use svn to checkout code for an open source project, i get this error: $ svn co svn://svn.valgrind.org/valgrind/trunk valgrind svn: Berkeley DB error for filesystem '/home/svn/repos/valgrind/db' while opening 'nodes' table: Cannot allocate memory svn: bdb: Lock table is out of available locker entries Can you please tell me how to fix it? Thank you.

    Read the article

  • Saving not cached offline images in Firefox

    - by Wroblewski
    I want to save the images from a page in Firefox that have been open for a couple of days. The problem is that neither the page nor the images are available anymore online. The images where dynamically generated and not cached. The only way I can think of is by making a screenshot and then cutting out each invidual image. Isn't there a better way of getting the data from the browsers memory?

    Read the article

  • Automatically copy files to USB drive when connected

    - by Daphna
    I am looking for a solution for copying all the files from a specific directory on the hard drive, to a specific directory on a USB memory device, once this device is connected. I have a program that downloads podcast episodes for me. I would like these files to be automatically moved (or at least copied) to my mp3 player once I connect it to the computer. I have both windows xp and linux machines, so a solution for any of them will work for me.

    Read the article

  • how can I monitor my FTP server [closed]

    - by hanem
    I have installed the FTP server in my PC and I have connected with Filezella client,but when I try to monitor the FTP server fonctionality by taping the snmpwalk command ,I don't have any result concerning the disk space of the server or the memory size,have any one an idea how can I monitor the FTP server and thanks the OS is windows7 and the SNMP is configured,I want to use the commandes of the net-SNMP library,but when I tape: snmpwalk -v1 -c public address of the FTP server I got the same informations as my system

    Read the article

  • Things to consider when building a continuous integration server?

    - by Dave
    I'm new to continuous integration, but immediately realize its value, and I want to get one set up right away. I have played with TeamCity and have it working in a VM great. Now, I don't want to spend money on another system, so I was planning on just doing the VM again on a faster machine (i.e. my dev system). There are a few questions that come to mind with this: Hard disk allocation - how big should it be? Sure, 60GB seems like more than enough, but people also used to think that we'd never need more than 64KB of RAM Backups - is it even important to back up the integration server? Sure, I guess it's nice so that one doesn't have to go through the entire configuration process again, but I would think that's about it. I could snapshot my VM every time I do a configuration change, and then do a backup of applications only (ignore the buildAgent stuff). Migration - if I want to go away from a VM on my dev system, to a new server, which maybe even runs Windows Server 2003, is it easy enough? Perhaps this is a particular point best suited for StackOverflow.

    Read the article

  • How to partition my two hard drives

    - by Thoma Bigueres
    I've got a computer running under the OS "Window Server 2008 R2" on which i have : 60GB disk C: NTFS (Disk 0) 40GB unallocated memory (Disk 1) I would like to partition my disk so that i'll have : 30GB disk C: 70GB disk D: Can you help me on the step i should do to be abble to have this configuration ? I saw that first of all i should merge the two volumes into one, but when i click right on the c: Volume, i can't click on the "Extend Volume" link. Do you know how i can overcome this ? Thanks a lot

    Read the article

  • Benchmark virtual machines?

    - by evan
    I'm looking for a good way to benchmark the performance of Ubuntu machines (preferably from the command line - only care about harddrive speed, memory, and cpu - not graphics). Are there any programs that could also be used on Mac or Windows so I could compare the results against an Apple or PC Desktop? Ultimately I'd like to use these benchmarks to compare different virtual machine configurations (speeds on different hosts and different hardware to get practical idea of the differences between different setups a rigs). Thanks!

    Read the article

  • How to reduce Windows XP computer boot time?

    - by Suma
    Are there any specific known steps which I could take to make my computer with Windows XP Professional booting faster? I am interested in speeding up following stages in particular: loading the OS (Windows logo, up to the moment login screen appears) log in user (from the moment you type your user name and password up to the moment all memory resident programs and services are loaded and the computer is really ready to use)

    Read the article

  • Is the XP VMM a bottleneck on a multi core machine?

    - by JeffV
    I have a dual Xeon hex core machine running an IO intensive application. (WinXP 32) I am seeing a hardware driver (1/2 user mode, 1/2 kernel, streaming data) that is using 6k delta page faults per second. When other applications load or allocate large amounts of memory the driver's hardware buffer gets an underrun (application not feeding it fast enough). Could this be because the kernel is only using one core to service page fault interrupts?

    Read the article

  • Server Hardware on the Desktop

    - by jcnnghm
    When I rebuild my desktop, I'm thinking of using server hardware instead of desktop hardware. I want to do this so I can easily add a lot of ECC memory (~20GB), and possibly more than one processor. I know that video hardware could be a problem, especially because I use 4 monitors. I should be fine with this, as long as I have two pci-e channels. Are there any downsides to doing this? Anything I'm not seeing?

    Read the article

  • slow windows 2008 server

    - by andytimmons
    my friend's company has a win2008 server, not R2, and it's really slow, it has 8G ram, has a SAP business one running on it, and it's also an AD, DHCP, DNS server, has Kaspersky 6 AV running as well. CPU usage is constantly 100%, physical memory is around 70%-90% even close everything, disable AV, if check processes, taskmgr.exe and windows explorer use like 40% each sometimes, do you have any suggestion what could be

    Read the article

< Previous Page | 419 420 421 422 423 424 425 426 427 428 429 430  | Next Page >