Search Results

Search found 18151 results on 727 pages for 'upside down'.

Page 371/727 | < Previous Page | 367 368 369 370 371 372 373 374 375 376 377 378  | Next Page >

  • Is it safe to use consumer MLC SSDs in a server?

    - by Zypher
    We (and by we I mean Jeff) are looking into the possibility of using Consumer MLC SSD disks in our backup data center. We want to try to keep costs down and usable space up - so the Intel X25-E's are pretty much out at about 700$ each and 64GB of capacity. What we are thinking of doing is to buy some of the lower end SSD's that offer more capacity at a lower price point. My boss doesn't think spending about 5k for disks in servers running out of the backup data center is worth the investment. Just how dangerous of an approach is this and what can be done to mitigate these dangers?

    Read the article

  • Recently reloaded computer, now have trailing effect when windows overlap

    - by Tony
    I recently did a clean install of my home machine and I am having a weird graphical error where if I move one window across the top of another it leaves a trail: I have the most recent NVIDIA drivers installed and nothing new is installed on my machine than before the reload (and I never had this issue before). I have two monitors and it happens on both so I'm sure it isn't a hardware issue (the card and my main monitor are also both less than 6 months old) I've also noticed a ripple when scrolling down on web pages which I believe is related. Any ideas what might be causing this and solutions to fix it?

    Read the article

  • Cheapest High Available Web Server [closed]

    - by xyz
    I would like to create a high-available setup (e.g. a small cluster) for a webserver, i.e. it will run Apache, PHP and MySQL. There will be between 2-8 small websites running with only very little traffic and workload. High availability is however very important. I don't want to be dependent on 1 datacenter, so there must be a minimum of 2 servers placed in different datacenters, and if one server goes down, the user must experience no or only a minimum of downtime - and no data loss. I have considered Amazon AWS using their Elastic Load Balancing, since it is possible to buy 2 EC2 instances in 2 availability zones and set up load balancing and RDS (Multi-AZ). However this seems rather expensive. Using the AWS price calculator http://calculator.s3.amazonaws.com/calc5.html it totals to 185$/month the first year (including the free tier). Are my calculations incorrect or is there a cheaper way to make this HA setup? Best regards

    Read the article

  • Ubuntu One, compressed files

    - by user8179
    I have uploaded some files to my Ubuntu One account and it seems to work great most of the time. I usually upload them directly from Nautilus by right clicking the folder, using the ”Synchronize this folder” option, and then I make sure that the file I want to upload is published. Then I usually test the whole thing by trying to download it. I right click the file again to get its URL and I paste it into my Web browser. This usually works fine. But yesterday I uploaded two compressed files – ”.tar.bz2”. When I tried to open them after downloading them with my Web browser (Opera), it failed. I found that the file was bigger than the original file (2358 B instead af 2335 B – 15 B added at the beginning of the file and 8 B added to the end), and someone at the Opera channel (IRC) at OperaNet (Europe) figured out that the reason for this is that the server compress the file again, ”without telling Opera”. So to be able to extract the file I need to add ”.gz” to the file name and then extract it twice. If I downloaded it with Firefox however, I didn't need to do that, so maybe Firefox figured this out somehow in a way that Opera does not. Someone also tried to download the file with wget and some other browser and he also got the same result as I did with Opera, that is the file is compressed a second time by the server. I guess ”the server” is the Ubuntu One server, right? So why is this? Could it be done better somehow? Or did I do something wrong when uploading the files? It also seems like this extra compressing thing does not always happen, because when I tried again a few minutes ago, the file came down with its right size (2335 B), without an extra compression. But the other file (114 MiB) was still compressed twice.

    Read the article

  • load balancing two web servers each on two different isp's?

    - by Scott
    I have two ISP's that provide me hosting via apache / php / mysql. I am running drupal on them. On occasion the mysql server will go away (crash), so I was hoping to find a reasonable way to have a fail over, if server A SQL is down, all traffic is sent to server B. I know traditionally this is handled in DNS where a second alternate ip is given if there is a problem - or similar. But I do not have control over the isp, other than I can run php, perl and the usual apache stuff. Also, I have static ip's on each isp, and I can create dns entries (A/CNAME/TXT). So, I was hoping there might be a way for me to have a script that checks if drupal has a problem, and if so, somehow alter dns, or ? Or, any other ideas? (other than spending lots more $ on a better isp)

    Read the article

  • ESXi 5.1 - Unable to register host

    - by deanvz
    I download and successfully installed ESXi 5.1. I am however unable to get the licence key I received installed. An error occurred when assigning the specified licence key: The system Memory is not satisfied with the 32 GB of Maximum memory limit. Current with 80.00 GB of Memory. Is there now way around this? A quick google revealed that this is a global problem with no real answer or resolution. The only workaround is to remove the physical RAM chips, but as this is in going to be in production I dont want to do that as it would mean down time when I have to reinsert the memory

    Read the article

  • ArchBeat Link-o-Rama for December 12, 2012

    - by Bob Rhubart
    “Cloud Integration in Minutes” – True or False? | Bruce Tierney The answer is 'True, but..." according to Bruce Tierney. "Connecting on-premise and cloud applications “in minutes” is true…provided you only consider the connectivity subset of integration and have a small number of cloud integration touch points." Get the rest of the story in Bruce's detailed post. Tech World Discovers New Species: The Cloud Architect | Wired Enterprise | Wired.com This Wired article by Cade Metz boils down to one essential conclusion: Cloud computing is a significant departure from "data center designs of the past," and the demand for the specialized skills of the cloud architect will only increase. But you already knew that, right? Oracle B2B - Synchronous Request Reply | A-Team - SOA "Beginning with Oracle SOA Suite PS5 (11.1.1.6), B2B supports synchronous request reply over http using the b2b/syncreceiver servlet," says C. D. Wright of the Fusion Middleware A-Team. His post includes a demo and everything you need to run it. Thought for the Day "Don't worry about what anybody else is going to do… The best way to predict the future is to invent it." — Alan Kay (Month Day, Year - Month Day, Year) Source: SoftwareQuotes.com

    Read the article

  • Server downtime - are these APC warnings the cause?

    - by DisgruntledGoat
    Yesterday I had a problem with my dedicated server (Ubuntu 10.04, LAMP). It wasn't down per se, but running incredibly slowly as if we had a massive overload of visitors (though I don't think we did). It's running smoothly again now. I've been checking through log files etc to see if I can find any issues, the only strange thing is a bunch of these errors, occurring at about the same time as the downtime: [apc-warning] Unable to allocate memory for pool. in [file] on line 49. And a bit later on: [apc-warning] GC cache entry '[file1]' (dev=2056 ino=8988092) was on gc-list for 3601 seconds in [file2] on line 746. Could these errors indicate the cause of the server slowdown, or are they simply a result of the server being slow in the first place? What would be the solution?

    Read the article

  • JiglibX addition to existing project questions

    - by SomeXnaChump
    Got a very simple existing project, that basically contains a lot of cubes. Now I am wanting to add a physics system to it and JiglibX seemed like the simplest one with some tutorials out there. My main problem is that the physics don't seem to be working how I imagined, I expected my tower of cubes to come crashing down, but they dont seem to do anything. I think my problem is that my cubes do not inherit DrawableGameComponent, they are managed by a world object that will update and render them. So they are at no point put into the games component list. I am not sure if this means that JiglibX will not be able to interact with them as in all the tutorials there are no explicit calls to add the Body objects to the physics system, so I can only presume that they are using a static/singleton under the hood which automatically hooks in all things, or they use the game objects component list somehow. I also noticed that in alot of the tutorials they use the following when setting up the physics system: float timeStep = (float)gameTime.ElapsedGameTime.Ticks / TimeSpan.TicksPerSecond; PhysicsSystem.CurrentPhysicsSystem.Integrate(timeStep); Would it not be better to keep a local instance of the created PhysicsSystem object and just call myPhysicsSystem.Integrate(timeStep)?

    Read the article

  • Corosync - stopping the service crashes the server

    - by Antipop
    I am trying to set up a test cluster on a Xen Server with 2 paravirtualized CentOS 5.4 machines. I am using Pacemaker+Corosync, and following the instructions found at http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf and other sites. Anyway, when I try to manually stop the corosync service, about 80% of the times the whole VM locks up with the message "Waiting for corosync services to unload" and I am forced to shut the machine down manually. For the remaining 20%, the VM keeps responding and adds dots to the above message, but it won't actually stop the service. There aren't many resources on the internet about this particular error. Any ideas about this? Thanks in advance.

    Read the article

  • Mail being delivered to both old and new server.

    - by Duane
    Two days ago I migrated our old site to a new server. I turned down the ttl before I did it and some email is still going to our old server and some to our new after a little over 48 hours. Both these are on shared hostgator sites. From what I can tell everything is set up properly. An nslookup points to the right ip address. Any insight? Support just tells me that everything is set up properly and I have to wait for propagation. I've never had it take over 48 hours though.

    Read the article

  • How to ask developers to think about high resolution screens

    - by WhiteWind
    I just got a ASUS N56VZ laptop, and it`s all good, but the screen resolution 1920x1080px. I am not asking, how to scale ui elements. If someone is interested I have found some tricks: increase font size in system settings increase unity dock size in MyUnity or system settings in modern Ubuntu versions. tweak userChrome.js of FireFox to make buttons|panels|icons larger add DefaultZoomLevel extension to FireFox to make it zoom pages initially. But all of it is miserable, because there are some big bugs: window decoration elements are way too small to pick them with mouse. I can't scale window easily and I can't position my cursor fast on the close|maximize buttons. Tuning lines like in sound volume dialog are hardly clickable at all. Unity top panel (status panel & tray) hardly can contain the bigger font, so it looks ugly, but icons are still the same. Sometimes I can`t read text, as it is cropped (and I cant scale some dialogs as it has fixed size) Chromium is not usable at all (ok, it's not Ubuntu problem, but the problem still exists) JAVA applications are not scalable (same as above) In FF I am able to get descent results in most cases, but multiplication of system font increase and browser font increase makes system controls (combobox, lists, drop down lists) extremely big, so I cant even control the zoom level on the page. IMHO, we should post a bug report (but what kind of bug?) and vote for it! The problem is even deeper, but at least we should ask developers to think about it. So my question is: how can I post a report (the right words and right place) and how can we (who already has that problem, or who want it to be solved before hardware upgrade) vote for faster solution. Any ideas?

    Read the article

  • Best VoIP VPS Location?

    - by ToiletOverflow
    Our office currently runs a few phones over a VoIP line. Through our VoIP provider, we have a virtual private server. We chose them as the VPS provider because the VPS came pre-loaded with all of the software that was necessary. However, I've discovered that I would rather manage the software myself and would prefer to work on a different platform. The primary reason that we have stayed with them is because as our VoIP provider, I presume that there is some advantage in call quality because they have "direct access to the PSTN", which I would presume is an advantage when it comes to call termination and overall call quality. My question boils down to: What is better from a call quality perspective? 1) A server located 20ms closer to us (60ms), offered by a different company. 2) The current server at the VoIP/SIP provider (80ms)

    Read the article

  • Highly Available Web Application (LAMP)

    - by Anthony Rizzo
    I work for a small company who provides a web application for thousands of users. Earlier this year they had one server hosted one company. We recently acquired another server in a different location with the hopes of one day making this a redundant failover machine. I understand what to do with the mysql replication, I plan on using a master-master replication setup, and rsync to sync the scripts and files, however I am at a stand still about how to configure the fail-over. Ideally I would like the two machines to accept requests, like a round robin dns, however if one machine goes down I do not want requests to go that machine. All of the solutions I am come across assumes high availability of servers in the same location, these servers are in two completely different locations with different public ip address. Any help would be great. Thanks

    Read the article

  • Mouse Hang with voltage fluctuation

    - by user265495
    I bought a new assembled PC about a month ago, but for the past 7 or 8 days I found my mouse cursor hangs every time there is a voltage fluctuation on my UPS (I guess it as when sound comes from ups then mouse hang for 2 or 3 sec. and them it goes normal again). My PC also seems to shut down while I'm working and it refuses to boot again - I end up needing to switch off my UPS and press and hold the power button for 30 seconds. It will then boot as per normal. Could the trouble be with my PSU? Am I potentially damaging my computer by pressing and holding the power button? I have : Motheboard : intel Chipset Model - DH61WWAAG23116-303 Processor : Intel Core i3 3.20 Ghz RAM : Zion 4 gb DDR3 HDD : 1 tb GRAPHICS : 2 gb NVIDIA GeForce GT-610 CABINET : Frontech Rock OS : Win 7 32 Bit

    Read the article

  • Dropbox fails to install on Ubuntu 13.10

    - by Fraxav
    I've done some research and it seems Dropbox has some problems with the status icon on Ubuntu 13.10, but I can't even manage to install it. The very first time, I installed nautilus-dropbox from the app center on a freshly installed Saucy, but I couldn't see the status icon and clicking the app icon wouldn't help, so I uninstalled and rebooted. The second and third time, the app just freezed at about 75% while installing from app manager, so I had to shut down the computer and uninstall. Next I tried to install dropbox from the .deb (from website): it installed (the launcher?) from app manager, then the dialog prompted me to start dropbox and finally to download the official daemon, but the download freezed at 0% and made everything else buggy/laggy (probably also because I restarted nautilus, as required, at that point). After a purge and clean from terminal, some icons and this file (which I suspect is somehow responsible, but of course is locked) are still present on my computer. Can anyone help me figure how how to install it? How could I remove that configuration file so I can give it a last shot?

    Read the article

  • Wifi won't connect to certain networks

    - by Dave M G
    As I write this, I'm in a coffee shop on my laptop, and I am trying to connect to the free wifi network they offer. Every time I select their network from the wifi applet drop down menu, it connects for a second and then disconnects. At the same time, I have my Android phone with me, and it connects to the coffee shop's network without problem. There are a couple of people around me also using the same network with no issues. I asked the staff to check and as far as they can tell there is no problem with their router or anything. I have tried rebooting, but it has not made a difference. I have double and triple checked that I am using the correct WEP password. As an added wrinkle, when I connect to another network, it seems to connect (although I can't use them because they are paid gateway services). So I'm able to ask this question now because I have tethered my laptop to my phone, but it's slow and a workaround, not a solution. All indications are that it's a problem with my computer against this one network. Why would that be and how can I fix it?

    Read the article

  • How to limit router bandwidth?

    - by David
    Hello, Is there a way to configure my (D-Link DIR-615) router to throttle down the allowed bandwidth after a certain amount of bandwidth has been used? For instance, I want my router to operate normally up to 20GB. After 20GB I want the router to limit bandwidth to a fraction of the normal speed (perhaps 1/5th). I live in Canada, so in about a month, everyone is going to be billed based on the amount they used (usage based billing). Instead of the unlimited bandwidth that I am enjoying now, most people will be capped at 25GB and will have to fork out $2/GB of over usage. Thank you in advance for the help.

    Read the article

  • Auto-scaling EC2 Servers and Updating Code

    - by jstats
    We've come to the point where we need to set up autoscaling for our web server and I'm unsure how to go about the process of scaling servers and updating the the existing code without remaking a new AMI and changing the autoscale config to use it. I've read a bit about people bundling the new code and uploading it to s3 and having new servers grab the bundle on boot up but that doesn't seem all that pleasant either. Currently the web app's files live in a git repo, and when we update the code, we push it to github, ssh into the web app and run a hook to bring down the latest code. So I was thinking that another option could be to just run that hook on an hourly or daily cron task. Unfortunately that doesn't cover everything type of update (for example new blog posts' images and such which aren't included in the git repo) but it's something. Could anyone provide some advice on what a common solution is or anything as to why my proposed solution is a bad idea? Thanks all

    Read the article

  • How to determine non-movable files in Windows 7?

    - by David
    Is there a way to determine which unmovable files are preventing Shrink Volume from releasing the full potential free space? Background: I have a 90 GB partition with Windows 7 on it, and 60 GB free space. I want to shrink it down to about 40 GB, and use the reclaimed 50 GB for a separate data partition. The Shrink Volume tool in Disk Management is only willing to give me 8 GB back. My understanding is that this is because of immovable files. I've followed the instructions found here, which involved disabling hibernation, pagefile, system restore, kernal dump, making sure all related files were deleted, and defrag'ing. I have successfully followed those same instructions before on this same drive, and partitoned the original 150 GB space into 90 GB and 60 GB, but I'm not so lucky this time.

    Read the article

  • Is There a Cloud Over OpenWorld?

    - by Tony Berk
    If you have been to OpenWorld in the past, you know it can be overwhelming or at least a bit "large." If this is your first time at OpenWorld, get ready! You are in for a big (or I should say HUGE) treat. The first thing you'll notice when you get to San Francisco is there are a lot of people, buses with "Oracle" posters, large exhibit halls filled with demos, games and tchotchkes from vendors with hot new solutions, and then there are the sessions. Yes, in fact there are over 2000 sessions. How can you possibly sort through 2000 sessions to find the best 20 or so for you? Which are the 1% for you? We will try to help with some insight over the next few weeks.  I'm going start at the highest level. Up in the Clouds! Since I know many people are looking for an update on The Oracle Cloud. We will drill down into the cloud and other topics for CRM and Customer Experience sessions in the next set of posts. Below is a list of some of the Oracle executive keynotes during OpenWorld highlighting The Oracle Cloud and applications related topics (the full list is here). In these sessions you will get details on Oracle's strategy and how Oracle is changing the industry to help our customers be more efficient, effective and innovative. Sunday, September 30 6:00pm - 7:00pm Larry Ellison: Hardware and Software, Engineered to Work Together: Why it's a Different Approach Tuesday, October 2 8:45am - 9:45am Thomas Kurian: The Oracle Cloud: Oracle's Cloud Platform and Application's Strategy Tuesday, October 2 3:30pm - 4:30pm Larry Ellison: The Oracle Cloud: Where Social is Built in Thursday, October 4 9:45am - 10:45am Mark Hurd: See More, Act Faster: Oracle Business Analytics We encourage you to also join the keynotes on the Oracle Database and Cloud Infrastructure and the fascinating partner keynotes, as well. Check the full list on the OpenWorld site. Oh, if you haven't registered yet, what are you waiting for? OpenWorld Registration Details.

    Read the article

  • Video problem with Windows 7 Games

    - by Jordan 1GT
    I have a Dell xps M1330 which originally ran Vista, but I upgraded to Windows 7. When I try to run a Win 7 game like spider solitaire I receive the following message: "The game is running in software rendering mode. Hardware acceleration is either disabled or not supported by your video card driver which could slow down game performance. Make sure you have the latest video card driver installed and that hardware acceleration is turned on." I confirmed that hardware acceleration is turned on. When I go to Dell's site, I'm told there is no later video driver. When I run the game it runs very choppy. I wouldn't care, but I loaded a .pptx file which is doing strange things in normal view and I suspect may be related to the same video problem. Any ideas?

    Read the article

  • Can't select text with mouse in Word / Office 2007

    - by asc99c
    I'm having a very weird problem here for the last few months. In Word, and in fact all programs from the Office 2007 suite, I can't drag the mouse pointer to select text. I can click at a point in the text and the cursor moves correctly to that point. If I double click, the word under the cursor is selected, and triple clicking selects the whole line. However if I hold the mouse button down and drag the mouse, no text is selected. Occasionally the problem disappears and everything works fine, but it then reappears a few minutes later. Text selection with the mouse works everywhere else (Firefox, PuTTY, OpenOffice), just not in Office. The only addins are Google Desktop Office Addin, and Person Name (). For info it is Office 2007 SP3, running on Windows 7 64-bit.

    Read the article

  • My partition table is facing limitations; having problems creating more partitions

    - by Terence
    I first noticed an issue when trying to install Linux Mint 14 as a third OS alongside Ubuntu 12.10 and Windows 7 - I was unable to create another partition to install Mint to. Poking around, I realised that I had reached the limit of primary partitions: (from left to right of the table) 1) a ~100 MB primary partition that I meant to use for storing Grub files but never got down to, 2) a 25 GB extended partition, with the entire partition containing an ext4 Ubuntu partition, 3) a ~513 GB partition containing Windows 7, 4) some ~50 GB of unallocated space, and finally 5) a 4 GB swap partition. I decided to wipe off the first ~100 MB partition, which I didn't need anyway. This brought me to a dilemma, however: as Ubuntu, during initial installation, had created an extended partition for itself, which meant that I could not create any more. Having an extended partition for Ubuntu is virtually useless as there is only enough space to contain Ubuntu itself, and the benefits of extended are worthless here. And if I go ahead and install Mint now, I would face the same issue of too many primary partitions in the future should I want to create any more. What I'd really like to do is this: convert the current extended partition into a normal primary partition, and create a new extended partition in the unallocated space for all future partitions. Is this possible, and if so how? Thanks in advance!

    Read the article

  • Dim (NEARLY blank) laptop screen, secondary screen works - why?

    - by LIttle Ancient Forest Kami
    My laptop screen is (almost) black while my secondary screen is fine. I believe it to be backlight / brightness related. Problem description it starts when I start the laptop system loads and works fine, just screen has problems I can see the screen though very faintly / dimly - it's hard to see anything which ain't very white e.g. starting screen has big Thinkpad logo in white, large font - I can see it, though very dimly second screen works very well Official backligtht debugging: using acpi setting as prescribed there for Thinkpads didn't help I can see an entry in /sys/class/backlight/ and it changes when I press hotkeys for brightness (current backlight power for instance goes up or down) acpi-off didn't helpm neither did acpi_backlight=vendor Hardware data Laptop is Thinkpad Edge with glossy screen. 4 processors, 2 cores, exemplary CPU data from cat /proc/cpuinfo reports Genuine Intel i5 (M 480 @ 2.67GHz). OS is Ubuntu Lucid, 10.04 LTS, 64-bit, with Linux generic kernel (2.6.32-44) and GNOME 2.32.2 (though I doubt there lies the problem). $ lspci | grep VGA 01:00.0 VGA compatible controller: ATI Technologies Inc M92 [Mobility Radeon HD 4500 Series] $ lshw -C display *-display description: VGA compatible controller product: M92 [Mobility Radeon HD 4500 Series] vendor: ATI Technologies Inc physical id: 0 bus info: pci@0000:01:00.0 version: 00 width: 32 bits clock: 33MHz capabilities: pm pciexpress msi bus_master cap_list rom configuration: driver=radeon latency=0 resources: irq:33 memory:c0000000-dfffffff(prefetchable) ioport:2000(size=256) memory:f0300000-f030ffff memory:f0320000-f033ffff(prefetchable) Driver I was NOT running any proprietary drivers, just checked with "Hardware drivers". There is one for ATI that is suggested there, though I didn't need it so far. UPDATE: changing the driver to proprietary one (ATI/AMD FGLRX) didn't help. Tried and failed Resetting / running on power or battery / charging / getting rid of static electricity / warming up *doesn't help* This is NOT a blank-screen problem, at least it isn't following official Ubuntu black-screen diagnostics - I can see my screen, though barely. What I will try next: - check last updates I've made - IIRC I am running on nomodeset already, but will verify this Any ideas how to proceed best? What is most probable cause?

    Read the article

< Previous Page | 367 368 369 370 371 372 373 374 375 376 377 378  | Next Page >