Search Results

Search found 10285 results on 412 pages for 'cpu architecture'.

Page 267/412 | < Previous Page | 263 264 265 266 267 268 269 270 271 272 273 274  | Next Page >

  • oracle virtualbox doesn't work for host Kubuntu since Lucid Lynx 10.04

    - by 13east
    I have a thinkpad edge 14 core i3 2.4 ghz and 4g ram; I have tried kubuntu 10.04, 10.10, 11.10, 12.04 and 12.10 (all x64 architecture); Both oracle and ose virtualbox only works properly to install XP and windows 7 quest system on kubuntu 10.04; For every other kubuntu release since, the guest installation goes as far as formatting the virtual drive, freezes at this step, and doesn't even go as far as copying files to hard-drive to begin installation. But virtualbox has not stopped responding to commands; I can kill that one specific window with the problem installation ("machine" - "close" - "power off the machine") and start over again without having to force-kill virtualbox application. If anyone knows how I can go about addressing this problem, any help you can provide would be very much appreciated. Thank You.

    Read the article

  • Agile: User Stories for Machine Learning Project?

    - by benjismith
    I've just finished up with a prototype implementation of a supervised learning algorithm, automatically assigning categorical tags to all the items in our company database (roughly 5 million items). The results look good, and I've been given the go-ahead to plan the production implementation project. I've done this kind of work before, so I know how the functional components of the software. I need a collection of web crawlers to fetch data. I need to extract features from the crawled documents. Those documents need to be segregated into a "training set" and a "classification set", and feature-vectors need to be extracted from each document. Those feature vectors are self-organized into clusters, and the clusters are passed through a series of rebalancing operations. Etc etc etc etc. So I put together a plan, with about 30 unique development/deployment tasks, each with time estimates. The first stage of development -- ignoring some advanced features that we'd like to have in the long-term, but aren't high enough priority to make it into the development schedule yet -- is slated for about two months worth of work. (Keep in mind that I already have a working prototype, so the final implementation is significantly simpler than if the project was starting from scratch.) My manager said the plan looked good to him, but he asked if I could reorganize the tasks into user stories, for a few reasons: (1) our project management software is totally organized around user stories; (2) all of our scheduling is based on fitting entire user stories into sprints, rather than individually scheduling tasks; (3) other teams -- like the web developers -- have made great use of agile methodologies, and they've benefited from modelling all the software features as user stories. So I created a user story at the top level of the project: As a user of the system, I want to search for items by category, so that I can easily find the most relevant items within a huge, complex database. Or maybe a better top-level story for this feature would be: As a content editor, I want to automatically create categorical designations for the items in our database, so that customers can easily find high-value data within our huge, complex database. But that's not the real problem. The tricky part, for me, is figuring out how to create subordinate user stories for the rest of the machine learning architecture. Case in point... I know that the algorithm requires two major architectural subdivisions: (A) training, and (B) classification. And I know that the training portion of the architecture requires construction of a cluster-space. All the Agile Development literature I've read seems to indicate that a user story should be the "smallest possible implementation that provides any business value". And that makes a lot of sense when designing a piece of end-user software. Start small, and then incrementally add value when users demand additional functionality. But a cluster-space, in and of itself, provides zero business value. Nor does a crawler, or a feature-extractor. There's no business value (not for the end-user, or for any of the roles internal to the company) in a partial system. A trained cluster-space is only possible with the crawler and feature extractor, and only relevant if we also develop an accompanying classifier. I suppose it would be possible to create user stories where the subordinate components of the system act as the users in the stories: As a supervised-learning cluster-space construction routine, I want to consume data from a feature extractor, so that I can exist. But that seems really weird. What benefit does it provide me as the developer (or our users, or any other stakeholders, for that matter) to model my user stories like that? Although the main story can be easily divided along architectural-component boundaries (crawler, trainer, classifier, etc), I can't think of any useful decomposition from a user's perspective. What do you guys think? How do you plan Agile user stories for sophisticated, indivisible, non-user-facing components?

    Read the article

  • How do I tell if a Python pip install is working correctly or hanging?

    - by wobbily_col
    I am trying to install the python pandas package in a virtualenv using pip. On my development machine it installed correctly, but now I am trying on the server, it gets so far then it seems to get stuck: warnings.warn(LapackSrcNotFoundError.__doc__) /apps/PYTHON/2.7.3/lib/python2.7/distutils/dist.py:267: UserWarning: Unknown distribution option: 'define_macros' warnings.warn(msg) non-existing path in 'numpy/distutils': 'site.cfg' non-existing path in 'numpy/lib': 'benchmarks' Could not locate executable gfortran Could not locate executable f95 Found executable /apps/modules/wrappers/fortran/ifort Top shows ifort running at 46% cpu. Is there any way I can tell if this is working correctly (can I check files it is updating for example), or if it is stuck in a loop? It has been running for 40 minutes so far.

    Read the article

  • Why has my Mac been running fsck_hfs for two days now?

    - by Nate
    I first noticed that fsck_hfs was running, taking up 50-75% of a CPU, yesterday. It continues to run today. ps shows that it is doing /sbin/fsck_hfs -f -n -x -E /dev/disk3. Only problem: I don’t think I have a /dev/disk3. Why is it running? Will it ever finish? Can I kill it? What is /dev/disk3? Could it be my Time Machine volume, which is not mounted at the moment? System Info: MacBook Pro (2008). It has two disks installed—the internal disk (/dev/disk1) and a PC Card SSD (/dev/disk0, surprisingly). It connects to a remote Time Machine volume attached to an Airport Extreme base station.

    Read the article

  • How do I troubleshoot a server I can't get logged into?

    - by Karl
    This morning I am unable to log into SQL Server because it appears to be busy doing something, so much so that CPU usage is constantly at 98%-99%. I am guessing that something has gone wrong with the maintenance plan that ran over the weekend, but I can't get into the error logs see what is happening. On logging in I keep getting messages like "insufficiently memory in the buffer pool" and "the server is not responding". Is there any way I can figure out what is happening without logging onto to the server? Or some way I can squeeze in a login?

    Read the article

  • Windows 7 x64 Hard Freezing (again)

    - by Lanissum
    Awhile ago, my computer was randomly freezing a few minutes after booting, and I ended up replacing the CPU and mobo after testing the RAM and hard drive, I also couldn't find anything wrong with the video card. So after replacing the presumably faulty hardware, everything worked fine for about a month and a half. All of a sudden, My computer is randomly freezing a few minutes after loading up any intensive application (games, mostly). Most of the time it just freezes with the current frame until I hard reset, although once it printed a BSOD message stating that dxgmms1.sys was to blame. The only difference between these two episodes I can think of is that I can do word/internet/work without issue now, as opposed to the near uselessness my computer was rendered last time. For those of you who want to know, I tested my memory with memtest86 (for 64 bit machines). I can't figure out what could have started this latest round of issues, the event logger just states that a kernel-power event has occurred (like last time) but I think thats just a generic "this machine has rebooted after a sudden shutdown" message.

    Read the article

  • Is a PCIE X1 card for USB 3.0 worth it?

    - by lelouch
    I have a portable HDD on USB 3.0, and soon I'll have a 32gb flash drive on USB 3.0 as well. My motherboard is pretty old - and the next time I'll upgrade it is when I get a new CPU/mobo/ram. Slots on my mobo: PCIE X16 - For my GTX 260 2xPCI PCIE X1 I'm considering purchasing a PCI express X1 card like this - http://www.ebay.com.au/itm/PC-Laptop-2-USB3-0-Ports-PCI-E-PCI-Express-X1-Card-New-/280772000446?pt=AU_Networking&hash=item415f50babe Now the question: What are realistic (read/write) speeds I'll get in this setup, assuming I have 1 USB 3 device connected at a time, and the bottleneck is the PCIE?

    Read the article

  • SQL server availability issue: large query stops other connections from connecting

    - by Carlos
    I've got a high-spec (multicore, RAID) server running MS SQL 2008, with several databases on it. I have a low throughput process that periodically needs a small amount of information from one of the DBs, and the code seems to work fine. However, sometimes when one of my colleagues does a huge query against one of the other DBs, I see full CPU usage on the machine, and connections from my app time out. Why does this happen? I would have thought the many cores and harddisks would somehow (together with cleverly written DB server) be able to keep at least some of the resources free for other apps? I'm pretty sure he doesn't use multiple connections for his query. What can I do to prevent this?

    Read the article

  • Monitor status while using VNC

    - by kumar
    So after connecting to a vnc server via vnc viewer to my desktop (remotely), Is it possible to know whether the monitor connected to the CPU is switched ON or not. Simply put, from command prompt how do you know whether monitor is ON/OFF from command line. Here, Basically I am bit worried about privacy as my monitor can be viewed by anyone while accessed remotely. Any solution? Obviously there is a option to switch off the monitor while starting the vnc server at remote side but I am looking for a better solution to control monitor(possible??) remotely. Thanks!

    Read the article

  • Xen Disk Performence Issues

    - by user98651
    I'm currently using Xen PV on CentOS 5 with my domU's as flat files running on a hardware RAID controlled (write cache enabled) formatted with XFS. On the dom0 I can get about 500MB/s in a 2GB dd write from /dev/zero however on the domU's I'm lucky if I get 10MB/s (it is usually around half that). I've tried changing the disk scheduling to NOOP on the domU's, changed some mount parameters and tweaked the performance allocations of both the dom0 (prioritize CPU) and domU's (increase RAM and VCPU allocations). None of these steps have produced any noticeable change in performance. My instinct here is that it is not a hardware problem, due to the solid performance of the dom0. Any ideas on what might be causing this problem? I'm considering moving to LVM based domU's.

    Read the article

  • nagiosgraph new services not showing

    - by Eleven-Two
    I am using Nagios Core with Nagiosgraph and had only enabled graphing for cpu usage for a while. This worked fine, but now i wanted to add some more services (for example memory usage). The new services are not working (no rrd data is generated). The Nagiosgraph site only says "no data available" and I get no error in apache log, nagiosgraph.log or nagiosgraph-cgi.log. The new services are standard services (nsclient++ MEMUSE for example) and of course they are included in the map file. If I execute the checks manually, it shows also the perfdata. I added the services by enabling the "graphed-service" use. Did I miss something?

    Read the article

  • Cloning Windows 2003 Server to new hard drive results in failure

    - by Level1Coder
    Scenario: Old hdd is a Seagate 320gb SATA drive New hdd is a WD 320gb SATA drive Created an exact clone and replaced old hdd with new hdd. Boot up with new hdd, it gets into Windows 2003 server environment but things look weird. Lots of system event failures in the event viewer log. System is barely unusable, critical services are all down. Boot up with old hdd, everything is fine. QUESTION: Is it possible to do a simple clone of a Windows 2003 server system? All I'm changing is the hard drive, everything else stays the same (old CPU/old mobo/etc..)

    Read the article

  • One bigger Virtual Machine distributed across many Nodes [on hold]

    - by flyer
    I just setup virtual machines on one hardware with Vagrant (this is just a test environment, not production!). I want to use a Puppet to configure them and next try to setup OpenStack. I am not sure If I am understanding how this should look at the end. Is it possible to have below architecture with OpenStack after all where I will run one Virtual Machine with Linux? ------------------------------- | VM | ------------------------------- | NOVA | NOVA | NOVA | ------------------------------- | OpenStack | ------------------------------- | Node | Node | Node | ------------------------------- (In my environment Nodes are just virtual machines, but my question concerns separate Hardware nodes) After some comments... Is it a language barrier, or? This is only my 'virtual environment'. If we imagine this virtual machines are a separate Nodes (e.g. every has 4 cores) the OpenStack is still the same, right? Can I run one Virtual Machine across many Nodes with OpenStack? Is it possible to aggregate the computation power of separate machines in one virtual distributed operating system?

    Read the article

  • New power supply and now HDDs are not recognized

    - by Michael
    So I upgraded to a new X4 ULTRA power supply that was recommended to me by a local TigerDirect store. After installing it along with a new liquid cooling system, I booted it up and it automatically fried my CD Drive. After that I noticed that the OS wouldn't start and figured out that none of the 4 HDDs in my computer were being recognized by the BIOS. I feel them spool at a steady pace and have tried new cables and connections but to no avail. I triple checked all of the connections and cables and have no idea what is wrong. This isn't the first time I changed a PS or CPU cooling system but I am at a dead end. Any ideas, aside from buying a USB HDD reader and seeing if they are all fried? Also, this is a stock Gateway mobo with the mobo USB connections already dead. Could the new PS have fried the SATA connections??

    Read the article

  • I just deleted "/bin". What's the best way to recover?

    - by Tom Marthenal
    I just ran (not on purpose!) rm -rf /bin. I've booted down the computer and am using Finnix to try to recover from it. I have succeeded in mounting the drive, and confirmed that, yes, the entire /bin folder is deleted. Is it possible to recover from this without reinstalling the OS? I'm thinking that I could setup a VM with the same OS and architecture (Ubuntu Server 11.10 alpha release, x86) and install all the packages I had installed on the server, then just copy the /bin folder. Will this work? Am I better off just starting over?

    Read the article

  • How do I know if my disks are being hit with too many I/O reads or writes or both?

    - by Mark F
    I know a bit about disk I/O and bottlenecks relating to this especially when relating to databases. How do I really know what the max I/O numbers will be for my disks? What metric might be available to me for working out roughly (but needs to be a good approximation) of how much capacity (if you will) have I got left available in I/O. I've seen it before where things are bubbling along nicely and then all of a sudden, everything screams to a halt, and it ends up being an I/O bound problem. Is there a better way to predict when I/O is reaching its limits? This article was interesting but not giving the answer I desire. So, is my best bet surrounding just looking at 'CPU I/O WAIT'? There must be a more reactive method than this.

    Read the article

  • Very Hot P4 3.4 GHz

    - by Timothy R. Butler
    My Pentium 4 3.4 GHz is running hot even with a new fan on my case's heatsink (80mm rated at ~ 40 CFM). If I stress the CPU, it goes over Intel's maximum temperature rating. I'm contemplating putting in a "PCI cooler" (i.e. a fan that fits into an empty PCI slot) to see if I could get better airflow in the case. The slot isn't far from the processor and heatpipes that lead from the processor to the fan, so I'm hopeful this would help, but not entirely certain. Thoughts?

    Read the article

  • High frequency, kernel bypass vs tuning kernels?

    - by Keith
    I often hear tales about High Frequency shops using network cards which do kernel bypass. However, I also often hear about them using operating systems where they "tune" the kernel. If they are bypassing the kernel, do they need to tune the kernel? Is it a case of they do both because whilst the network packets will bypass the kernel due to the card, there is still all the other stuff going on which tuning the kernel would help? So in other words, they use both approaches, one is just to speed up network activity and the other makes the OS generally more responsive/faster? I ask because a friend of mine who works within this industry once said they don't really bother with kernel tuning anymore-because they use kernel bypass network cards? This didn't make too much sense as I thought you would always want a faster kernel for all the CPU-offloaded calculations.

    Read the article

  • Linux IO scheduler on databases with RAID

    - by Raghu
    Hi, I have a linux database(MySQL) server(Dell 2950) with a 6-disk RAID 10. The default IO scheduler on it is CFQ. However, from what I have read and heard, there is no need for a scheduler like CFQ when reordering/scheduling is also done by underlying RAID controller; on the contrary since it does not account underlying RAID configuration into account performance may actually degrade with CFQ. The primary concern is to reduce CPU usage and improve throughput. Also, I have seen recommendations of using noop/deadline IO scheduler for databases primarily because of the nature of their R/W access.

    Read the article

  • What is a proper MySql replication configuration for frequent db updates and rare selects?

    - by serg555
    We currently have 1 master db on its own server and slave db on app server. App executes very frequent but light updates (like increasing counters), and occasional (once in a few minutes) heavy selects (which is the most important part of the app). When app was connected only to master db there were no performance issues. With slave db introduction CPU load avg on app server increased to about 6-10 during that heavy select period (from 3-4 as before). When server doesn't run those frequent updates it seems like performance for selects stays within the limits. So I have a feeling that those updates is what is causing the performance drop (also these frequent updates are not critical so if slave db doesn't have them in sync with master for some time it would be ok). What would be a good db replication setup for such kind of app? What are the replication parameters we could tweak? Thanks.

    Read the article

  • HDD not detected whenever first power on (detected at BIOS level)

    - by Alvin Lim
    I am using Asus P8H61-M LX Motherboard with latest version of BIOS: 4401 One SATA-3 Western Digital Black Series 1.5TB is attached to SATA port 0 A Samsung 22x DVDRW SATA drive attached to SATA Port 1. ps/2 Logitech mouse and keyboard. CPU is i3 with DDR3 8Gb RAM. no other devices are installed. My problem is that when I turn on the computer, the WD HDD is not detected. ie cannot be booted. I have to press ctrl+alt+del in order to get it detected. I'd used the WD data lifeguard software to scan the HDD, the smart status is OK, the surface scan does not show any error at all. Where did I do wrong? Any advice is greatly appreciated.

    Read the article

  • how to debug BSOD irql_not_less_or_equal error

    - by Sev
    So I'm getting the irql_not_less_or_equal BSOD. I tried looking at the event viewer for potential causes, and cannot find any. I also checked the CPU temperature in BIOS right after the error and that was fine. I've tried 2 sets of RAM chips already, both give the issues. The error doesn't happen consistently...it happens daily, but many hours can pass and it won't happen, or only 10 minutes can pass and it might happen again. By the way, just bought the parts and built the computer myself a couple of weeks ago. How to debug the cause for this? Hardware info: Asus P6X58D PREMIUM motherboard Intel core i7 930 quad core 2.8 ghz Kingston 128 GB SSD 3 Gb/sec Nvidia Geforce GTX 465 PNY Edition Corsair 12 GB DDR3 1600 Mhz Ram Windows 7 Ultimate

    Read the article

  • PC Speaker sounds when machine locks up.

    - by d.c
    What would cause my PC speaker to sound a low frequency continuous tone as my system locks up? Generally what will happen is I'll notice the application I'm using will stop responding but explorer will be partially responsive, then after a few more actions explorer stops responding, and the last click will cause the pc speaker to emit a low constant tone as the machine locks up completely. The only resolve at that point is to restart. Its not a thermal issue, I'm reading the cpu, and hdd temps with software and hardware monitors. AV and malware scans come up clean. I've swapped out my ram, reseated all my components. Used sfc with no results. chkdisk locks up at 3% and defragmenting does the same, but I can read the drive without trouble (I know this as I've done av/malware scans and I also backed everything up since this started happening) I'm mostly just interested to learn, if anyone knows, why the pc speaker would sound during the lock up. windows xp sp2

    Read the article

  • HAProxy being killed with more that 54,000 connections

    - by Olly
    I am trying to run HAProxy (1.4.8) on a EC2 machine running Ubuntu 10.04. I need HAProxy to be able to handle many thousands of long-running persistent connections (websockets). With the current setup HAProxy gets killed at around 54,300 connections (roughly). If I am running HAProxy in the foreground, the only output is "Killed". Am I right in thinking this is the Kernel killing the process? Is this because it is out of resources? Can I increase the resources? The CPU and memory consumption are low with 50,000 connections, so I don't suspect either of these. How can I prevent this from happening?

    Read the article

  • Firefox hangs with java in Linux

    - by Xolve
    I am using Firefox 3.6 on Linux and I haveJava 6 update 15. Obviously I am using the newer plugin type. Whenever I open a page with Java applet it just becoms unresponsive and hangs. I tried with Chrome. I think it detects the Java plugin from mozilla's plugins folder. Due to multiprocess architecture it doesn't hang and shows the initial applet loading window but then it reports that the plugin has crashed. The only way I can run Java applets is to download the class file and make an html file and run the applet using applet viewer.

    Read the article

< Previous Page | 263 264 265 266 267 268 269 270 271 272 273 274  | Next Page >