Search Results

Search found 15099 results on 604 pages for 'runtime environment'.

Page 351/604 | < Previous Page | 347 348 349 350 351 352 353 354 355 356 357 358  | Next Page >

  • Accessing multiple local HD's or RAID with ESXi 4.0

    - by Shawn Anderson
    How to I get additional HD's to be recognized and used by ESXi 4.0. When I purchased my system I had two 2TB HD's, but when I installed ESXi it only recognized one of them. I'm happy to get whatever number of drives that I need (I have a four bay SATA in my Dell T310). What are some options? RAID? If so, is it supported. I guess I would need hardware instead of software since ESXi is so small. The VMWare forums (where I've lived for the last two days) are a charlie foxtrot of outdated and conflicting info. I want to utilize my T310, with 32 GB RAM, 2.8GHz quad core to run many lab Windows VM's. I don't need production level availability but I do want decent performance, even though it's in a lab environment. A huge thanks to Jim B., Zypher, Helvick, and Jeff Hengesbach who posted answers to my earlier predecessor question on why ESXi was so sluggish.

    Read the article

  • Better to develop Ruby project on server or buy a faster desktop computer? B/c laptop too slow.

    - by user33184
    I have a Linux laptop (Vostro V13) running a Celeron M chip. This is a fine laptop, but running unit tests especially for Rails applications is slow. I want a faster development environment but I don't want to spend too much money. So the choice I have is between $390 for a Linux desktop machine with a Pentium Dual Core Processor E5400 and just paying between $30 and $40 a month to Linode and trying to do development remotely on that server. Can anyone with experience developing server applications using both methods offer any advice?

    Read the article

  • Forwarding emails to nonexisting users/aliases to external mail server

    - by Niclas Lindqvist
    I'm in the works of installing a postfix mailserver on a machine currently being used as web server. As of right now, I've got it working as far as that I can send and receive emails using telnet through port 25. However, as my customer is concerned with downtime, I'd like to setup all the accounts one by one over time, and making sure it works, rather than just cutting the cord to the old mail server and start creating new accounts on the new system. How can I add the domain customer.com to my mail server and just add something like [email protected] to the users and aliases-lists without the web server trying to send all emails with the @customer.com domain to the new mail server? I'm running ubuntu, using postfix and postfix-mysql on the new machine, the old mail server is on a hosted environment somewhere else, where I don't have any control.

    Read the article

  • Options for PCI-DSS on AWS - file integrity monitoring and intrusion detection

    - by Brill Pappin
    I need to deploy some file integrity monitoring and intrusion detections software on AWS instances. I really wanted to use OSSEC, however it does not work well in an environment where servers can auto deploy and shut down based on load, because it requires server managed keys to be generated. Including the agent in the AMI will not allow monitoring as soon as it comes up because of that. There are many options out there, and several are listed in other posts on this site, however none that I've seen so far deal with the unique problems inherent in AWS or cloud based deployments in general. Can anyone point me at some products, preferably open source, that we might use to cover those portions of PCI DSS that require this software? Has anyone else achieved this on AWS?

    Read the article

  • Cannot start Oracle XE 11gR2 Net Listener and Database on Ubuntu 13.04

    - by hydrology
    I have been following the setup step on this article for installing Oracle XE 11g R2 on Ubuntu 13.04. The environment variables PATH, ORACLE_HOME, ORACLE_SID, NLS_LANG ORACLE_BASE have all been set up correctly. simongao:~ 06:16:38$ echo $PATH /usr/lib/lightdm/lightdm:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/home/simongao/adt-bundle-linux-x86_64-20130219/sdk/platform-tools:/u01/app/oracle/product/11.2.0/xe/bin simongao:~ 06:18:36$ echo $ORACLE_HOME /u01/app/oracle/product/11.2.0/xe simongao:~ 06:23:29$ echo $ORACLE_SID XE simongao:~ 06:23:35$ echo $ORACLE_BASE /u01/app/oracle simongao:~ 06:23:37$ sudo echo $LD_LIBRARY_PATH /u01/app/oracle/product/11.2.0/xe/lib simongao:~ 06:23:48$ echo $NLS_LANG /u01/app/oracle/product/11.2.0/xe/bin/nls_lang.sh However, when I try to startup the service, I receive the following error information. simongao:~ 06:18:40$ sudo service oracle-xe start Starting Oracle Net Listener. Starting Oracle Database 11g Express Edition instance. Failed to start Oracle Net Listener using /u01/app/oracle/product/11.2.0/xe/bin/tnslsnr and Oracle Express Database using /u01/app/oracle/product/11.2.0/xe/bin/sqlplus.

    Read the article

  • Is there a way to create a consistent snapshot/SnapMirror across multiple volumes?

    - by Tomer Gabel
    We use a NetApp FAS 6-series filer with an application that spans multiple volumes. For backup purposes I would like to create a consistent snapshot that spans these volumes at the same point in time (or at least with an extremely low delta); additionally, we'd like to to use SnapMirror to replace the production environment to test volumes. The problem is in creating a consistent snapshot/SnapMirror, since these commands are not transactional and do not take multiple parameters. I tried scripting consecutive "snap create" or "snapmirror resync" commands via SSH, but there's always a 0.5-2 second difference between each snapshot. It's currently "good enough", but I'm seriously concerned about the consistency impact with increased load (we're currently in pre-production). Has anyone managed to create a consistent snapshot that spans several volumes? How did you pull it off?

    Read the article

  • Looking for a free or open-source burner emulator [closed]

    - by Jared Harley
    Possible Duplicate: Virtual CDR driver I am looking for a free or open-source virtual CD/DVD emulator to run in a Windows environment. What I want is similar to what SlySoft's Virtual Clone Drive or Daemon Tools provides, but the emulated drive needs to be a burner of some type. The burner should be able to save disc images (.iso, .ccd, etc) to my harddrive - basically, the same as if I burned the files to a CD-R, and then ripped them back to a disc image. I have already looked around some and come across 2 - DVD neXt COPY iTurns and NoteBurner M4P. Both of these programs create a virtual CD-RW drive, but they are integrated into their product (for burning from iTunes to create mp3 files) and cannot create disc images. I am currently writing a piece of software that will have the capability to burn disc images onto CDs/DVDs, and I don't want to end up with a 100 coasters while I'm testing my software. Anyone have any ideas? Related ServerFault queston: Create netbook recovery image without DVD burner (virtual burner?)

    Read the article

  • disable RADIUS for Cisco 2500 wireless controller

    - by Tim Vaughan
    I have a Cisco 2500 wireless controller and four lightweight access points. I want to use the controller to manage a wireless network secured by WPA only, without using RADIUS or anything else. We'll handle the authentication using a captive portal behind the access points. However, it seems like the controller's default security policy requires a RADIUS server and I can't find out how to switch the policy off. The documentation assumes I'm in an environment which needs heavy-duty security and the use case is actually a small charity/business with much less stringent security requirements. How do I disable the complicated security policy and instead run a simple one that just uses WPA?

    Read the article

  • Outlook 2010: Cached Exchange Mode, File Storage and Security

    - by dangowans
    I'm in an environment where profile space is a premium, and most users have "frozen" machines, meaning that on restart, the C: drive is returned to its original state. Cached Exchange Mode sounds interesting to me, but I'm wondering if we can take advantage of it without causing other issues. Where in the file system does the cached data get stored? Is it in the profile? A temp folder? Is the cached file secured in some way to keep others from seeing it?

    Read the article

  • Exchange 2007 Email Address Policies

    - by Ryan Migita
    We have recently upgraded to Exchange 2007 (from 2003) and have noticed the change from recipient policies to email address policies. We have two separate domains (let's call them domaina.com and domainb.com) we receive email for, have email address policies and both email address policies are not applied. In our Exchange 2003 environment, domaina.com was the default email address when we created new mailboxes and due to the migration, domainb is the default (and its email address policy is a higher priority). Now, when we create a new mailbox (or edit existing ones), the primary email address becomes domainb.com. Now the question is, is this as simple as putting the email address policies in the correct order? Do I have to apply both policies? What effect will the above changes make to existing mailboxes? Since we do not have any conditions set on the policies, I assume prior to making these changes, I should force all domainb mailboxes to not automatically update email address based on policy? Thanks in advance!

    Read the article

  • How can I access an Apple Xserve with no I/O ports

    - by DigitalJedi805
    I have an Apple Xserve at my place of employment, that I stumbled across one day going through some old equipment. The Xserve has one card installed on it, being the NIC, and is labelled with the static IP adresses that it has assigned, but other than that it is totally 'headless'. I would like to put it to use ( by either rolling it over to a Linux, or Windows Server 08 environment... 0_o ) but have yet to figure out how to get into the system to manage it. I'm a frequent at StackOverflow, but this is my first SuperUser post, so please let me know if this should be on one of the affiliate sites.

    Read the article

  • Why do all my sites randomly stop in iis?

    - by Jen
    I have a GoDaddy Account with a virtual hosting environment. For some reason every few weeks, all of our sites (both websites and FTP sites) are stopped. I restart them and they start normally, but I would like to prevent this from happening in the future. It is an iis thing. When you open up iis you can see that each of the sites has been stopped in iis. Also, I have tried looking at the event viewer and I don't see anything, but I am not sure what I am looking for.

    Read the article

  • /etc/profile.d and "ssh -t"

    - by petersohn
    I wanted to run a script on a remote machine. The simple solution is this: ssh remote1 some-script This works until the remote script doesn't want to connect to another remote machine (remote2) which requires interactive authentication, like tis one (remote2 is only reachable through remote1 in this case): ssh remote1 "ssh remote2 some-script" The solution for the problem is to use the -t option for ssh. ssh -t remote1 "ssh remote2 some-script" This works, but I get probems in case I use this (where some-script may execute further ssh commands): ssh -t remote1 some-script I found that some environment variables are not set which are set when I don't use the -t option. These envrionment variables are set in scripts from /etc/profile.d. I guess that these scripts are not run for some reason if using the -t option, but are run if I don't use it. What's the reason of this? Is there any way to work around it? I am using SUSE linux (version 10).

    Read the article

  • Get a file from a load balanced server in Windows Server

    - by Leandro
    I've a load balanced server on production environment for my application. The server is on Windows Server 2008 R2. I'm running a web application that creates and save a file into a folder on the web path. So I need to create a job that copy this file into another server. The main idea is that a file watcher checks for the file and then copy it instantly. But how can I know in what server it's the file? Please avoid "why you don't" answers to get a directly answer, if it's someone.

    Read the article

  • phpredis + pconnect

    - by john smith
    I am using phpredis on my php based website. The webserver I am using is a the simplest apache apt-get installation, no configuration involved, as this is only a development environment. The issue I am facing is that basically, while using phpredis, there is no difference between "connect" and "pconnect" commands: they both create a new connection everytime, as I can see from the "info" command on redis-cli. Now, I am pretty sure it is because of the apache configuration and the fact that it probably (most likely) is a multi-threaded env, therefore can't enstablish a single connection. My question is basically for when I will turn into production: what would the best choice of a webserver be to avoid this problem? I remember using lighttpd with thousands of users and still get only a very few (like 2 or 3) connections on mongoDB. Any ideas? Thanks in advance.

    Read the article

  • How to use wget with an input file and filenames

    - by Matt
    i have a text file that contains 10,000 url's with a unique number i want to save the file as. Each line has a 10 character code, then the URL of the image to retrieve. How can I make the input file use the first 10 characters as the wget filename? this is an example of the input file: input.txt x100083590http://image.allmusic.com/13/adg/cov200/drt200/t291/t29123q8m19.jpg b200149548http://ecx.images-amazon.com/images/I/41DoH%2BAWKEL.jpg z100151855http://image.allmusic.com/13/amg/cov200/dri400/i450/i45035hxdrb.jpg p400171646http://ecx.images-amazon.com/images/I/61cH4n34IhL.jpg wget -i input.txt would get the file but not with the preceding unique number. I want t29123q8m19.jpg (the first line) to be saved as x100083590.jpg If there is a better way to write out the input file, say with the URL first, then I can do that too, but I will never know the length of the first field. Right now the first 10 characters will always be what I want to save the wget image as. Edit This is being done in a windows environment.

    Read the article

  • Script or Utility to convert .nab to .csv without importing double entries in Outlook.

    - by Chris
    Currently our environment is migrating from Groupwise 7 to Outlook 2003 and we have multiple users with mission critical outside contacts in their frequent contacts that will have to be imported in Outlook. Currently our only solution is to export GW contacts to a .nab, import to excel to scrub out the contacts in our own domain (to avoid double entry) and convert to .csv. This current solution will require a lot of man hours for hand holding because most of our users are not technically savvy AT ALL and are frankly too busy to do this themselves. Anyone know of any kind of tool or script to assist with this?

    Read the article

  • Booting Windows7 kernel from an initrd/wim image file

    - by Ivo
    I'm wondering if it's possibile to have Win7 kernel and relative drivers (especially storage drivers) to boot from an initrd-like image file (maybe .wim?) and later then mount the windows root partition and complete the load of the full OS? I'll try to explain why: I'm running an emulated environment with NO REAL BIOS, and I'm passingthrough a raid storage controller. I want windows to boot from this controller array, but of course the BCD manager cannot access disks in the array until kernel and relative controller storage drivers are loaded. To be clear I get the classical winload.exe missing error. I need a similar solution to what Linux does, loading the kernel and his drivers, and later then mount the root partition and complete the boot. Any ideas or advices?

    Read the article

  • Network access to VM only from host ...

    - by Jamie
    I'm trying to do some testing of Ubuntu 10.04 Beta 2 Server in a VMWare environment. The host is XP, and the VM software is VMPlayer. Problem is, I want to be able to see the VM from the network, not just from the host. I can SSH into the VM from the host, but from any other machine on the network I can't even get the VM to respond to a ping. Going the other way (from the VM out) isn't a problem at all. The VM software did an 'easy installer' so I'm not really sure what was going on with the networking. Suggestions?

    Read the article

  • putty pageant - forget keys after period of inactivity

    - by pQd
    in the environment where windows client computers are used to run putty to connect to multiple linux servers i'm considering moving away from password based authentication and using public/private key pairs with pass-phrases. using ssh-agent would be nice, but at the same time i'd like it to 'forget' the pass-phrases after given period of inactivity. it seems that putty's pageant does not provide such feature; what would you suggest as alternative? solutions that i'm considering: patching pageant code [might be tricky, code is probably quite rusty and project - sadly - stagnant] writing small custom application using GetLastInputInfo and killing pageant if the machine was idle for more than let's say 15 minutes [ yes, there'll be separate policy for locking the desktops as well ] using alternative ssh client and ssh agent. any suggestions? thanks!

    Read the article

  • Minimum space required for Server 2008 R1 install?

    - by Chris J
    I'm trying to plan disk space for a virtual environment, and wanting to keep virtual disks as small as possible - mostly as apart from the base OS, the software going onto the VM is less than a few MB, so want to avoid physical disk space going to waste; plus it'll give me an idea of how many VMs I could physically fit in {x}GB of physical drive. For Server 2003, I've had installs on 2GB and 5GB sized virtual disks. However for Server 2008, Microsoft recommends a minimum of 10GB (I assume this is both for x32 and x64). For the record, I will be installing the x32 version. Now I know I could just go ahead and try a small install, but wanted to solicit any practical knowledge as well :-) What's the smallest install of Server 2008 possible? (excluding server core installations).

    Read the article

  • Is it a good idea to have the operating system on a solid state drive?

    - by Kenji Kina
    There is something I don't quite understand. I know a SSD helps with OS load times, but I'm not sure if all this boost is only noticeable/interesting when booting, or gives an all around considerably better experience thereafter. I am interested in having a quick and responsive environment after booting, which leads me to think that it'd be better to spend the SSD capacity in my most used apps (and the page file? Another inside question) and not the OS itself. This, of course, means that I don't know just how much the OS reads/writes its files during normal usage. So, how good an idea is it to dump the whole 20GB+ of Windows 7 OS into the SSD (considering the hefty price per GB of SSD capacity) if I can put up with the usual hard disk boot times? Would I be missing on a lot if I didn't?

    Read the article

  • Secure Server Distro

    - by Drama
    Hello, I have a root-server (i7/24GB/1TB) running Ubuntu 10.04 LTS as my OS. After some security audits (OpenVAS, Retina etc) I see that Ubuntu isn't the most secure system for a semi-corporate environment. Its updated from many sources, ofc from the Ubuntu security repo too. But nevertheless I could exploit my OpenSSL install with an exploit from August/September. There are some critical updates needed which Ubuntu does not provide. I was using Debian and Ubuntu for almost 5 years but now I doubt. What distro is secure and up to date from your point of view? How can I make the server more secure? Outsourcing of every software-module to a VM? I am not new to server-hardening, my packages are up to date I read Ubuntu Security Notices and I have no unneeded services installed on my server. Thanks.

    Read the article

  • Python and mod_wsgi path issue

    - by jasonh
    I have an AIX 6.1 system that I've compiled and installed: Apache 2.2.21 (into /usr/local/mercurial) Python 2.7.2 (into /usr/local/bin and /usr/local/lib) mod_wsgi 3.3 (with the AIX fix #1 described here) Mercurial 2.0 (system-wide) However, when Apache starts, I get the following message in error_log: IOError: invalid Python installation: unable to open /usr/local/bin/lib/python2.7/config/Makefile (No such file or directory) See the problem? bin/lib doesn't exist. /usr/local/lib/python2.7/config/Makefile does exist though. However, I can't figure out where it's getting that path from. Here's the environment variables I've got: PYTHONHOME=/usr/local/bin PYTHONPATH=/usr/local/lib/python2.7 LIBPATH="/usr/local/mercurial/lib:$LIBPATH" PATH=/usr/local/bin:/usr/local/lib:$PATH LDR_CNTRL="MAXDATA=0x80000000" AIXTHREAD_SCOPE=S AIXTHREAD_MUTEX_DEBUG=OFF AIXTHREAD_RWLOCK_DEBUG=OFF AIXTHREAD_COND_DEBUG=OFF SPINLOOPTIME=1000 YIELDLOOPTIME=8 MALLOCMULTIHEAP=considersize,heaps:8 I've tried all sorts of combinations with and without PYTHONHOME, PYTHONLIB and PATH in envvars. My PATH, in case it matters is: /usr/bin:/etc:/usr/sbin:/usr/ucb:/usr/bin/X11:/sbin:/usr/opt/ifor/bin:/usr/local/bin:.

    Read the article

  • Puppet: Conditional file source based on naming convention

    - by thinice
    I'm getting the ball rolling on puppet for my environment - and I'd like to have a conditional file resource based on whether or not the module itself contains a file based on a naming convention. So visually, assume a module named 'mysql' and it's layout: mysql/ /files /etc/ my.cnf my.hostname1.cnf my.hostname2.cnf /manifests init.pp ... So I'd like the block to verify if the resource for the module exists or not, and take action accordingly, in pseudo-terms: file { '/etc/my.cnf': if -f 'puppet:///mysql/etc/my.$hostname.cnf' { source => 'puppet:///mysql/etc/my.$hostname.cnf' } else { source => 'puppet:///mysql/etc/my.cnf' } } This way one wouldn't have to manage a csv file or the .pp file with a host specific case statement - is this possible?

    Read the article

< Previous Page | 347 348 349 350 351 352 353 354 355 356 357 358  | Next Page >