Search Results

Search found 11759 results on 471 pages for 'isolation level'.

Page 234/471 | < Previous Page | 230 231 232 233 234 235 236 237 238 239 240 241  | Next Page >

  • 504 Gateway Time-out after php fatal error

    - by tiagojsag
    I'm using nginx and php-fpm to develop a symfony2 based website, under ubuntu 12.10 (yes, I know I'm using a beta OS). Everything was working out fine until, due to an error on my code, I called an unexisting function, and got the following: Fatal error: Call to a member function (....) This isn't a problem (it's a bug in my code, easily fixable), but after this, no other page loads. My browser just keeps trying to load the page from the webserver, until nginx timeouts (after +- 30s, which should be some default timeout) and returns: 504 Gateway Time-out Restarting php-fpm solves the issue. Nginx logs show a timeout message, and nothing appears on php-fpm logs, even if I set them to debug level. I tried switching from fpm to fastcgi, and the same thing happens. I've looked around, but all similar error are related to big requests/file handling, which isn't the case. All the pages on my website load in a few seconds, even under development conditions (no caching, etc).

    Read the article

  • How secure is Remote Desktop from OSX to Windows Server 2003?

    - by dwhsix
    It's unclear to me exactly how secure Remote Desktop access from OSX to a Windows Server 2003 machine is. Is the communication encrypted by default? What level of encryption? Are there best practices for making this as secure as possible? I found http://www.mobydisk.com/techres/securing_remote_desktop.html but it's unclear how much of that is still relevant for current versions of RDP and Windows Server. I know I can tunnel RDP over ssh, but is that overkill or redundant? Thanks...

    Read the article

  • Dell PowerEdge 6850 Degraded HDD

    - by Matt
    Good Morning, We have a dell power edge 6850 with a degraded drive in the RAID array. I have never had to recover such an issue, so any help or advice would be welcome. Basically it hasn't affected the server at an operating system level, but has slowed down performance, I have a replacement drive in hand but as this is our main database server I want to proceed with extreme caution. My options as I see them are - Can I just hot swap the degraded drive with the new one and the data will automatically re-sync and we are all back to normal presumably this is dependant on the current raid configuration? reading various comments on-line I may need to re-configure the RAID array and re-build the broken drive? This screams disaster to me with the main worry being that I wipe any other data. Option 1 would of course make my day. Thanks in advance

    Read the article

  • Lightweight, low cost enterprise backup solution

    - by Scott
    Looking for a backup solution primarily for Windows clients (XP/7), that will either back up to 2 different servers (1 on site, 1 off site - internet - can be our own server), or back up to 1 server and then we would need to somehow backup that server offsite/internet. By lightweight, I mean the backup client software should not eat up much memory and processor since some of the client machines are older. I am used to using Crashplan for home use - the pricing is nice for the amount of backup I get, and it works great / easy to install and get going - I can back up to my own machines locally and over the net. However, the price is going to be a little steep for enterprise level backup, 1500+ machines. Possibly ZManda and Bacula are good choices to consider? Are they light weight? Can the clients/agents be set to go over the net and/or multiple backup servers?

    Read the article

  • Possible to use DRBD on two ESXi virtualized servers?

    - by chen
    I have two servers (attached disks have been set up as hardware RAID1 for disk device level failure resilience). Here is the setup in my mind: 1) Install ESXi on each of the physical server, M1, M2; 2) Start one VM on each of the ESXi virtualized physical server V1, V2; 3) Install the DRDB drivers within V1 and V2. Essentially, this is a "virtualizing machine running DRBD in the VM's instead of bare metal hardware" idea. My question is whether the above setup can achieve the same "networked RAID1" goal that DRDB can achieve in the bare-metal physical machines (http://www.drbd.org/). Thanks. [EDIT] I found this (http://serverfault.com/questions/49305/drbd-experimentation-and-virtualization) is a similar question, but the answer does not seem to be firmative enough for me to follow.

    Read the article

  • DHCPPloc.exe or equivalent for Windows 7?

    - by Bart B
    There seems to be some DHCP funnyness going on so I need to run something to show me what's going at a DHCP level. Before I upgraded my machine to Windows 7 I used DHCPloc.exe from the Windows XP support tools, and it worked like a charm. I can't seem to find Support Tools for Windows 7, and trying to use the XP tools in compatibility mode doesn't work (I tried, it fails to open a receiving socket). I need a tool to monitor DHCP traffic, and ideally one that lets me filter it to exclude DHCP traffic from our trusted DHCP servers and only show me un-authorised DHCP traffic.

    Read the article

  • Xen and Ubuntu?

    - by wag2639
    How does one properly approach having Ubuntu servers on a Xen hypervisor? I don't have any experience with RAID or Xen other than from a theoretical level. Additional requirements: Use with mdadm Software RAID 5 (can be on separate disks) that multiple instances with access Paravirtualized Ubuntu Server guests instances Possible ideas for now: Ubuntu host (dom0) with ubuntu-xen-server package (this purportedly isn't supported) dom0 host will "own" RAID 5 partition More Ubuntu servers as guests Citrix XenServer bare-metal host XenServer will own RAID Citrix XenServer bare-metal host Ubuntu guest instance creates and owns RAID Questions and concerns: Can Ubuntu be used as a dom0 Xen host? Can XenServer install packages such as mdadm and create a partition? Can multiple guest access (R + W) to the same data partition (RAID)? Note: since it may have a bearing on support, I'm referring to Ubuntu Server 10.04

    Read the article

  • Best method(s) to backup VMs running on HyperV?

    - by Kara Marfia
    We're in the middle of P2V'ing most of the network, so the current backup method is likely the worst - the backup agent is still installed on the guest OSs, and the backup device is dutifully pulling them onto tape, one file at a time. I suspect there's a clever way to script (PowerShell?) a suspend on the VMs, then backup of the .vhd files, and unsuspend the VMs. This seems like it would provide big speed benefits, while losing file-level restore (might be best for things like DCs and app servers). What methods/policies have you hammered out?

    Read the article

  • Unable to remotely schedule tasks from the command line

    - by Eptin
    I'm on a Windows 7 machine, attempting to use the command line to schedule a task on another Windows 7 machine in my company's network. I have administrative-level credentials for both computers. With help from http://msdn.microsoft.com/en-us/library/windows/desktop/bb736357.aspx I have created this line to run on the command prompt: schtasks /Create /S machinename /U username /P password /SC ONCE /TN Test1 /TR C:\Windows\System32\calc.exe /ST 16:30 Whenever I launch that, I get the following error: ERROR: User credentials are not allowed on the local machine. How can I fix this?

    Read the article

  • 8 Character Device names

    - by Lee Harrison
    Is there any reason to still use only 8 characters in a device name? My boss still uses this rule for printers, computers, routers, servers... basicly any device connected to our network. This leads to massive confusion among users, especially when it comes to printer. It also leads to confusion from an administration standpoint because every device is named vaguely, and similarly(its only 8 characters!). I understand the history behind this and compatibility with older systems, but none of our legacy systems will ever make use of PS-printers and Wifi networks. Is there any reason to still do this, and what is everyone else doing when it comes to naming network devices at an enterprise level?

    Read the article

  • How to remove duplicate illegal site in apache configuration?

    - by zladuric
    I've recently found a referrer in the apache log on my site. Now, I opened it out of curiosity, since my site is live, but I just started development so I didn't expect it. Anyway, the site was a pure copy of mine, and after investigation I saw that it resolves to my IP. I'm on Ubuntu 12.04, Apache 2, drupal 7, don't know what other info can I provide. My question is: how can I tell apache that it should not serve this site? Thanks Edit: forgot to say that I had some bots register to my fresh drupal installation. Also, my domain is a tld, this fake domain is a third level (ie. sub.domain.de)

    Read the article

  • Software Engineer's explanation of Facebook photo privacy, with UML diagram?

    - by Alex R
    Facebook photo privacy is more complex than most people think - including the bloggers who fill the Internet trying to explain it in simple terms. Not only there is the basic album-level privacy setting to consider, but also what happens with Tagging (and related privacy settings) as well as the Share button when clicked by a Friend. Has anybody seen a good, engineering-type (e.g. UML) diagram? I envision it should include the various privacy "states" a photo can be in, what causes state transitions, and the characteristics of each state? Thanks

    Read the article

  • Resilient Linux Mail Server Setup

    - by Coops
    How would people design a resilient mail server setup with Linux? On an application level what the system needs to provide is both an incoming and outgoing mail service (i.e. SMTP & IMAP), along with filtering and archive storage (the archive part isn't critical yet, so we'll look at this later probably). What is required on top of this is a resilient system, i.e. one which will handle individual server failures without interrupting service. As such I would term this a High Availability mail system. This is in contrast to a High Performance mail setup, as in our case the volume of mail being handled isn't the important factor, it's simply that it stays online. Having not approached this problem before, the first thing I thought of was a clustered file system (gfs/gluster/etc), combined with heartbeat to failover a floating IP to another box in the case of a server failure. Combined with postfix & dovecot does this sound feasible to people?

    Read the article

  • Text template or tool for documentation of computer configurations

    - by mjustin
    I regularly write and update technical documentation which will be used to set up a new virtual machine, or to have a lookup for system dependencies in networks with around 20-50 (server-side) computers. At the moment I use OpenOffice Writer with text tables, and create one document per intranet domain. To improve this documentation, I would like to collect some examples to identify areas where my documents can be improved, regarding general structure and content, to make it easy to read and use not only for me but also for technical staff, helpdesk etc. Are there simple text templates (for example for OpenOffice Writer) or tools (maybe database-driven) for structured documentation of a computer configuration? Such a template / tool should provide required and optional configuration sections, like 'operating system', 'installed services', 'mapped network drives', 'scheduled tasks', 'remote servers', 'logon user account', 'firewall settings', 'hard disk size' ... It is not so much low-level hardware docs but more infrastructure / integration information in these documents (no BIOS settings, MAC addresses).

    Read the article

  • How do I analyze an Apache Bench result?

    - by Alan Hoffmeister
    I need some help with analyzing a log from Apache Bench: Benchmarking texteli.com (be patient) Completed 100 requests Completed 200 requests Completed 300 requests Completed 400 requests Completed 500 requests Completed 600 requests Completed 700 requests Completed 800 requests Completed 900 requests Completed 1000 requests Finished 1000 requests Server Software: Server Hostname: texteli.com Server Port: 80 Document Path: /4f84b59c557eb79321000dfa Document Length: 13400 bytes Concurrency Level: 200 Time taken for tests: 37.030 seconds Complete requests: 1000 Failed requests: 0 Write errors: 0 Total transferred: 13524000 bytes HTML transferred: 13400000 bytes Requests per second: 27.01 [#/sec] (mean) Time per request: 7406.024 [ms] (mean) Time per request: 37.030 [ms] (mean, across all concurrent requests) Transfer rate: 356.66 [Kbytes/sec] received Connection Times (ms) min mean[+/-sd] median max Connect: 27 37 19.5 34 319 Processing: 80 6273 1673.7 6907 8987 Waiting: 47 3436 2085.2 3345 8856 Total: 115 6310 1675.8 6940 9022 Percentage of the requests served within a certain time (ms) 50% 6940 66% 6968 75% 6988 80% 7007 90% 7025 95% 7078 98% 8410 99% 8876 100% 9022 (longest request) What this results can tell me? Isn't 27 rps too slow?

    Read the article

  • Fedora Core 11 won't boot without a monitor

    - by feihtthief
    I have a P4 system that I installed Fedora 11 on. It will not boot without a monitor attached. The monitor can be off (not even have power plugged in), but must be attached. Without a monitor the hard disk thrashes around a bit like it's starting up services, but does not get to the point where I can ssh into the box. I have set the default runlevel to 3 and removed the rhgb entry from grub. Any suggestions welcome. Edit: I have already set the run-level to 3. The machine boots up fine with the monitor plugged in to the point where I can SSH into it. as soon as i unplug the monitor and reboot, it will not boot to that point.

    Read the article

  • Client side certificates in client browsers with unix server for management

    - by user146253
    We are currently running Unix dedicated servers for everything (Web cluster, database, FTP, batch, ...) except for a Microsoft Active Directory Certificate Services. The sole purpose of this Windows box is to provide client side certificates to our clients browsers. All our clients are required to install a client side certificate on order for them to be able to access our website. Is there an alternative in the Unix space? The purpose is to make sure only the approved hardware of an approved client can access our website. I'm open for any solution that provides me with this level of security. We are however talking about thousands of certified computers just so you can factor that in in a proposed solution. Optionally we would also like to be able to revoke access. With Regards.

    Read the article

  • Windows XP, USB-Stick and multiple Partitions

    - by Bobby
    Hello. I've got an USB-Stick with multiple Partitions on it (FAT32 (active), FAT32, Ext2 <-- that's another story) and it seems like that my Windows XP can only mount the first partition of the stick. If I try to mount the second one using the volume manager it tells me that I need to make it active and reboot...is it really that limited or am I just missing something here? Partitions: FAT32, System Rescue CD, bootable and active FAT32, some tools ext2, some data (I know that I need extra drivers etc., but that's not asked here. Edit (Solution): Thanks to the answer with the RMB (ReMoveable Bit) I was able to dig up a solution described at this site (Section: On flash drive only the first partition works). Basically, there's an Hitachi Driver available which filters the RMB on Driver-Level, which just needs to be a little modified to function with basically every USB-Stick. All you need to do is adding the "Device Instance ID" to the driver and then use this driver.

    Read the article

  • IDE Motherboard to Boot from a SATA 1T Hdd

    - by JavaMan
    I want to use my SATA HDD in my very old ASUS A7V8X-X motherboard (made in 2003). I intended to buy a cheap adapter for this. Say something similar to this one: http://www.cooldrives.com/satoidecofor.html But would there be any issue if I want to boot from this SATA drive? My impression is that these kind of adapters convert SATA signal and commands to PATA directly and such a low level conversion should be transparent to the motherboard and BIOS - in case the BIOS doesn't support SATA. And, for my motherboard, it sure doesn't. Does anyone ever used such a kind of adapter to support a boot drive? Any success sorry?

    Read the article

  • Updating windows xp hard drive controller from ubuntu live disk

    - by Joel
    The problem: When booting, I get a blue screen shortly after the Windows XP logo splash screen. The error code is 7b, and the second hex number is 0xC0000034. Based on this link (item 7) it appears the driver should be updated. Oddly, I made no changes to the drivers recently. I suspect it was something in a windows update or the newest upgrade of my antivirus (eset). But I digress. The BSOD makes me unable to boot into Windows at all, so I can't update the driver from there. I've run various bios-level diagnostics (including full surface scan) and the hd looks good. I'm also able to boot to an old ubuntu disk and read files from the hd. The question: Based on the above, it appears that I need to update the Windows hard drive controller from the ubuntu live disk. How do I do that?

    Read the article

  • What is the best filesystem for storing thousands of files in one dictionary-like id-blob structure?

    - by Ivan
    What filesystem best suits my needs? Thousands or even millions of files in one directory. Good (ext4 & ntfs level or close) reliability (incl. fault tolerance) and access speed. No directories actually needed, as well as descriptive names, just a dictionary-like structure of id-blob pairs is all I need. No links, attributes, and access control features needed. The purpose is a file storage where all the metadata (data describing all the facts about what the file actually contains and who can access it) is stored in a MySQL database. As far as I know common filesystems like NTFS and ext3/4 can go dead-slow if there are too many files placed in one directory - that's why I ask.

    Read the article

  • How to enable directory browsiing in IIS7?

    - by frankadelic
    How I enable directory browsing in IIS7? MS technet says this can be done in the IIS console: Open IIS Manager and navigate to the level you want to manage. In Features View, double-click Directory Browsing. In the Actions pane, click Enable if the Directory Browsing feature is disabled and you want to enable it. Or, click Disable if the Directory Browsing feature is enabled and you want to disable it. http://technet.microsoft.com/en-us/library/cc731109%28WS.10%29.aspx However, my IIS console doesn't have the Directory Browsing option mentioned in Step 2. How can this option be made available. Note, this is for a static HTML site, so I don't have any web.config or ASPX files.

    Read the article

  • Strange Intermittent Background Sound in Windows 7

    - by NoCarrier
    Ok, this is very strange. Recently, i noticed that every 10-15 seconds, there would be this faint, annoying duh-dum sound coming out of my speakers. (hard to describe. sounds somewhat like the sound windows makes when you unplug a USB device, but not as pronounced and much quieter). I closed every app and ended as many processes as I could, but the sound persists. I look at the volume mixer and sure enough, when the sound occurs, there is a little spike in the level under "System Sounds". I haven't installed any hardware or software in the last several weeks. This started recently - completely out of the blue. Does anyone have any insight?

    Read the article

  • Enterprise online backup providers

    - by PHLiGHT
    We've used Iron Mountain's LiveVault service but found that it was only good for file level backups. We liked how it backed up every 15 minutes. It doesn't support Exchange 2007-10 and the web interface was very poor. Who else is everyone using? The most notable names in online backup such as Mozy and Carbonite don't really seem suitable for larger companies. We have SQL, Exchange and Sharepoint servers and are looking to virtualize in the near future. Until then bare metal restore capability would be nice. We are currently using Backup Exec 12.5 but that can be so troublesome at times. We have about 2 TB of data. 1TB is archival data.

    Read the article

  • What is needed for 'Previous Versions' to be visible on the client OS?

    - by Zoredache
    I have servers with Shadow Copies enabled taking snapshots a couple times a day. From the server, if you look at the local devices you can see the Previous Versions being populated reliably. But from remote clients, the ability for an end-user to see the Previous Versions seems to be very hit-or-miss. For the sake of this question you can assume that all my clients are Windows 7 and the Servers are Windows Server 2008 R2. Is there an exhaustive list of everything that is required for end user to see Previous Versions? Are their any requirements for a certain level of share or filesystem permissions, other then read access? Does something need to be open on the firewall, other then what is already in-place for normal Windows networking?

    Read the article

< Previous Page | 230 231 232 233 234 235 236 237 238 239 240 241  | Next Page >