Search Results

Search found 15099 results on 604 pages for 'runtime environment'.

Page 345/604 | < Previous Page | 341 342 343 344 345 346 347 348 349 350 351 352  | Next Page >

  • MacBook Pro for Windows development via virtualization. Performance?

    - by webworm
    I am a Windows/Web developer by profession and I have been considering a MacBook Pro as a replacement for my current development machine. I am impressed by the build quality, the uni-body construction and performance specs of the MacBook Pro. I am specifically interested in the 13.3" MacBook Pro running Core 2 Duo 2.4 GHz processor with 4 GB RAM. What I am wondering is this ... what performance can I expect running SQL Server 2008, IIS, and Visual Studio 2010 within a virtual environment (VMWare Fusion and Windows 7) on the above mentioned MacBook Pro? I like the 13.3" model as the size is more portable, but am I expecting to much from a core 2 duo processor? Would I need to look at the next step up in MacBook Pro using the core i5 processor? Thanks!

    Read the article

  • pam_exec.so PAM module does not export variable PAM_USER as stated in the documentation

    - by davidparks21
    I'm trying to use the pam_exec.so PAM module to execute a script which needs to know the username/password coming from the application (OpenVPN in this case). I have a script that executes printenv >>afile, but I don't see all the environment variables that the man pages states that pam_exec.so exports (namely PAM_USER I think), I only see the following: PAM_SERVICE=openvpn PAM_TYPE=auth PWD=/usr/local/openvpn/bin SHLVL=1 A__z="*SHLVL I do successfully pick up the password off of STDIN and output it with this same script. But for the life of me I can't get the username. Any thoughts on what I should try next?

    Read the article

  • P2V - Cold Clone ISO

    - by jlehtinen
    I need to cold clone a physical box in a VMWare environment. What are people using for this these days? My preference is for VMWare's vConverter ISO, but it appears that this was discontinued. It's no longer available for download on their site from what I can tell (even under old versions). I found one guy who appears to have an ISO for version 3.0.3 of vConverter posted to his site for download, but I'm eternally skeptical about downloading these types of software from random strangers: http://thatcouldbeaproblem.com/?p=584 I also found some mention of using MOA, but I've never used this and have no idea on how effective it is as a vConverter replacement. http://www.sanbarrow.com/moa.html One other options seems to be using Acronis - booting off an Acronis disk to capture a .tib, then using a standard installation of vConverter to push the .tib to ESXi.

    Read the article

  • Virtualizing WindowsXP on Linux Fedora 12 for Photoshop....

    - by Rae
    I am running a Linux server from 2000, I have Fedora12 installed as the OS. I know you can virtualize Windows XP on the system to be able to run Windows programs, but the problem is I My server has only one 1333Mhz Pentium 3 processor and 1G of Ram. granted this is a pretty kick ass computer for the day, but lacking the second processor I'm afraid that I will not be able to run Windows virtualization to support Photoshop CS3. Is there any program out there that runs similar to Photoshop, like windows Digital Image suite 10, that will run smoothly in my Linux Fedora 12 environment? or can I virtualize windows and run the Windows digital image suite 10, without slowing or corrupting my system?

    Read the article

  • sudo ENV_KEEP not always preserving

    - by mafro
    When I run sudo -s, my environment is preserved. However, when running a simple sudo <command> it appears not to be preserved. The contents of my sudoers file: mafro@ip-10-xx-xx-250:~ > sudo cat /etc/sudoers.d/mafro Defaults env_reset Defaults env_keep += "HOME" mafro ALL=(ALL) NOPASSWD:ALL Using sudo -s, the ll alias is available: mafro@ip-10-xx-xx-250:~ > sudo -s root@ip-10-xx-xx-250:~ > ll total 8K drwxrwxr-x 2 mafro dev 4.0K Jun 9 23:59 bin drwxr-xr-x 20 mafro dev 4.0K Jun 9 23:59 dotfiles Using straight sudo, it is not: mafro@ip-10-xx-xx-250:~ > sudo ll sudo: ll: command not found What is happening here?

    Read the article

  • xinet vs iptables for port forwarding performance

    - by jamie.mccrindle
    I have a requirement to run a Java based web server on port 80. The options are: Web proxy (apache, nginx etc.) xinet iptables setuid The baseline would be running the app using setuid but I'd prefer not to for security reasons. Apache is too slow and nginx doesn't support keep-alives so new connections are made for every proxied request. xinet is easy to set up but creates a new process for every request which I've seen cause problems in a high performance environment. The last option is port forwarding with iptables but I have no experience of how fast it is. Of course, the ideal solution would be to do this on a dedicated hardware firewall / load balancer but that's not an option at present.

    Read the article

  • What is a good OS to mess around with Xen Dom0?

    - by earlz
    Hello, I'd like to experiment some with Xen. My problem is I do not know which OS to try to install it with. I use OpenBSD, but it can not run as Dom0 or DomU at this point. I use Arch Linux on my website's VPS, so I know it and like it. The support for Xen just doesn't seem to be there though. Can anyone suggest a nice (preferably lightweight, so not SUSE and such) Linux distro or *BSD to mess around with Xen in a non-production environment?

    Read the article

  • Crystal Reports 8 - Error 533 PESStartPrintJob

    - by Federico Giust
    We have on the company I work for an application built in Delphi V with Crystal Reports 8. We all know that Crystal & Delphi can be temperamental sometimes and the worst thing is, nearly no detail on the error. Also there is a big lack of documentation on the web about this. There's lot of people with similar issue but no solution. The error I'm talking about in particular is the one on the image below. This happens when trying to print any Crystal Reports on screen. This has happened sometimes in the clients computer and it was hard for us to replicate on our environment. Since it's an old version of Crystal it is very hard to find any helpful documentation to know the exact source of the problem.

    Read the article

  • Full speed internal switch bandwidth but per-port set external bandwidth?

    - by garg
    I am in an environment where all the machines are behind a switch that I don't have access to. Each ethernet wall port has limited bandwidth depending on how much has been paid for each port. The problem is that some people have 10Mbps connections and some have 100Mbps connections and this causes problems with local intranet file transfers and operating system/software deployments. Operating systems can take hours to be deployed if the machine is on 10mbps. Do you know if it is possible with most switches to set a rule that would limit bandwidth coming in/going out to an extranet, but keep full bandwidth if the packets are destined to go to a local machine? For example, the internet might be limited to 10Mbps, but internal servers would get gigabit speeds? Thanks

    Read the article

  • OSX 10.6 integration into NIS/netgroup/automount infrastructure

    - by mdpc
    I have an existing infrastructure where accounts are maintained under NIS (yp) with no local unix accounts. Also, all the standard maps including hosts, mail aliases, netgroups, etc...are maintained in this form. Extensive use of the UNIX/Linux automounter with items scattered over the network on NFS servers. There are NO ACLs on any local or shared files. All mail needs to use basically the nullclient sendmail configuration feeding into a different system. I now have a requirement to integrate an Apple OSX 10.6 system into this environment and make it run seamlessly. My initial reading and second-hand information seems to indicate that this may not be possible on the native OSX 10.6 system. I'm concerned. Any ideas as to how to accomplish this task and make everybody happy? Thanks PS: I have never used an Apple OSX system.

    Read the article

  • How to setup a django site with Cherokee, DynDNS and virtual_env?

    - by e-satis
    I have a django project running with the dev server, and would like to try run it in a production environment. I wanted to try Cherokee for a change, so I installed it. We don't have a domain name yet, so I set up a DynDNS looking like stuff.gotdns.org. It works fine, I can see the Cherokee welcome page (so red, I first believed I got an error :-p). I ran the wizard to create a new virtual server for Django. No everything is setup, but I have nothing. Still the default Cherokee welcome page. What should I do now if I want to go to "http://stuff.gotdns.org" and see my website? What should I do now if I next want to make it available only at "http://project.stuff.gotdns.org"? Important fact, I use virtual_env, so your can call Python directly, you have to activate it first.

    Read the article

  • Co-location vs. cloud hosting

    - by RainyRat
    We have a decent-sized server estate: several ProLiants, plus IBM BladeCentre and SANs running a VMWare environment. Due to an imminent premises move, we're not going to have enough space for a server room, so I've been looking at moving everything out to a colo. My boss is more keen on the idea of cloud-hosting everything, which would include a couple of high-traffic websites (about 9m pageviews/month), our Exchange sever (about a million clean emails sent/received monthly), as well as file/print/AD/all the ususal stuff. This doesn;t sound like a good idea to me, but I'm new to the ways of the cloud. Can anyone offer any advice?

    Read the article

  • Re-using port 443 for another service - is it possible?

    - by Donald Matheson
    The ultimate goal is to allow a remote data connection service to operate on port 443 on a the SQL server. The application accessing the connection is behind a firewall and it is because of the client's reluctance to open another port that I have been asked to try and get this working. The current environment is Windows 2003 R2 (SP2) and SQL Server 2005. IIS is not installed, but when I try and install the third party connection software (SequeLink) it won't as it reports something is still configured on/using port 443. Netstat does not show anything listening on the port and I've tried editing the system32\drivers\etc\services file removing any reference to port 443 and also using sc delete to delete the HTTP and HTTPFilter (HTTP SSL in services console) services to see if this would help. Rebooting after each change. What could still be using the port? Is what I'm trying even possible (I have my doubts but have to investigate every avenue)? Any help/pointers would be greatly appreciated.

    Read the article

  • Can unexpected power loss harm a Linux install?

    - by Johan Elmander
    I am developing an application on a Linux embedded board (runs Debian) e.g. Raspberry Pi, Beagle Board/Bone, or olimex. The boards works on an environment that the electricity is cut unexpectedly (it is far complicated to place PSU, etc.) and it would happen every day couple times. I wonder if the unexpected power cuts would cause crash/problem on the Linux Operation System? If it is something that I should worry, what would you suggest to prevent the damages on OS against the unexpected power cuts? PS. The application needs to writes some data to the storage medium (SD card), I think it would not be suitable to mount it as read-only.

    Read the article

  • Remotely managing Hyper-V VMs from Windows 8 Client

    - by Vazgen
    Currently, I have a core Hyper-V Server hosting VMs for a domain controller and several domain-joined VDI infrastructure servers. The VMs are connected in that domain environment, but the remote management of the physical Hyper-V Server is set up using the same WORKGROUP (as the Windows 8 client I'm managing from) This makes it cumbersome to manage the VMs hosted on the physical server from my remote management Windows 8 client because I can only connect to the physical Hyper-V server and not the individual VMs hosted inside. Can I make my set up more flexible by hosting a second domain controller in a VM hosted on my Windows 8 machine and switching my remote management set up to use the same domain through? Meaning ALL physical and virtual machines including the VDI infrastructure under the same domain? I'm new to this just looking for some suggestions.

    Read the article

  • Using 2-port LSI 2308-8e card to control 24 SAS HDDs

    - by GregC
    I would like to rely on a RAID-on-chip solution to control 24 SAS hard drives in a direct-attached environment. How would you approach this to get best bandwidth given that I'd like to spend less than $10,000 on the interconnect. I've read that LSI 2308 chip can easily handle 8-drive SSD RAID6 in hardware. I'd like to harness its power to control 24 SAS hard drives over an expander in an external enclosure. Currently I use an Infortrend S24S-G2240 external enclosure. It provides its own controller and expander. I'd like to use LSI 2308 controller for RAID6 somehow instead of the mystery controller in the enclosure. P.S. I tried to create SAS-expander as a tag, but my rep on this site is low.

    Read the article

  • Can not run ifconfig like commands via browser

    - by savruk
    Hi, Problem is I cannot run "ifconfig" or similar commands via browser. Environment: Programming language : python Server : lighttpd(CGI) , running on busybox. Well machine is really small and so I am really restricted. Tried techniques: chown every script to root. But there is no differences. Why? Because lighttpd runs under another user, I mean not under root. As it is not root, when I try to run script from browser it always calls the python file with its uid. So it makes it impossible to run "ifconfig eth0 192.168.2.123" like commands via web browser. I get "ifconfig: SIOCSIFADDR: Permission denied" error. What can I do? I do not have any sudoers file, so cannot modify sudo command. Well, I don't even have "sudo" command :) Thanks for your help

    Read the article

  • Providing high availability and failover using MySQL on EC2

    - by crb
    I would like to have a highly-available MySQL system, with automatic failover, running on Amazon EC2 instances. The standard approach to solving this is problem Heartbeat + DRBD, but I've found a lot of posts suggesting DRBD doesn't work on EC2, though none saying exactly why. Obviously, a serial heartbeat or distinct network is out of the question in the virtualised environment. It would also be good to have the different servers be in different availability zones, but we're getting into a much harder problem there. What are peoples' opinion on having a high uptime solution in "the cloud"?

    Read the article

  • What paths are guaranteed to exist on Windows Server 2008 R2?

    - by jpmc26
    What paths are guaranteed to exist on a Windows Server 2008 R2 instance? A client is requiring that some instructions specify exact paths in all cases. (The person executing said instructions is not supposed to have to decide on any path themselves, even when the path makes absolutely no difference.) So I need to know what paths I can rely on to be there. It's fine with me if they involve environment variables, but they need to be variables guaranteed to hold an existing path. (That is, no modification to a path that doesn't exist possible.) Or are there no guaranteed paths?

    Read the article

  • How can I terminate wisptis.exe in Windows XP SP2

    - by user9950
    I am working in a production environment. Basically I work on graphics oriented enviroment so I used to work on catia and windchill as well. When I am opening a part or diagram in Catia along with CNEXT .exe (catia service) wisptis.exe is starting and making the system halt and Catia will crash. So I need to restart the worksation or I need to close all applications then I need to open catia then again the same problem. So please tell me how to avoid this or stop the wisptis.exe service. If I want to work on catia I need to kill wisptis.exe service otherwise I cant work.

    Read the article

  • Why won't my Windows 7 KMS key work on my Server 2008 KMS server?

    - by Ryan Bolger
    Our Microsoft volume licensing site was recently updated to include our Windows 7 and Server 2008 R2 KMS keys. We have an existing KMS server running on Server 2008 (not R2). In an attempt to be proactive about supporting the new OSes in our environment, I unregistered the old KMS key with slmgr.vbs and tried registering the new key. The registration failed with Error 0xC004F050. The description for that error was "The Software Licensing Service reported that the product key is invalid." What's wrong? I've checked and double checked that for typos against what is listed on the Volume Licensing website.

    Read the article

  • use of [!NOTFOUND=return] in nsswitch.conf

    - by Chris Phillips
    Has anyone come across the use of this config for passwd and groups config in nsswitch.conf? Where I'm working I've been told it's been shown to help situations where a group exists both locally and in ldap which was causing issues for group memberships etc. However this config seems to totally mess up nscd which will be aware of the groups and all their members but will not flip the data around to say the user is a member of all it's remote groups. Initially it seems, given a fully available environment, to be exactly the same as [FOUND=return] which is an implict default between stages anyway. However apparently a lengthy ticket with Redhat resulted in the recommended use of that configuration.

    Read the article

  • Configuring Apache for multiple clients

    - by Chris_K
    Last week I had a question here about suexec / suphp but I tried to accomplish too much. I'm going to narrow the scope a bit and try again. I'd like to configure a LAMP server to host multiple clients. I'd like it to seem (from the client's viewpoint) just like any other shared hosting environment. Web sites in their home directory, no need to muck around with file ownerships to get pages served, etc. It would seem that a configuration that involves suexec and suphp is the way to go(?) I'm specifically looking for a current/modern guide on how to accomplish this (I'll be using CentOS if it matters) and I'm afraid I need more than a link to Apache docs. Are there any good How-To's out there? The few I've found have been pretty out of date, but it is quite possible my search was weak.

    Read the article

  • Best practices for re-IP'ing / migrating servers and applications

    - by warren
    Some of this question would be highly application-specific, but what approaches do you take when looking to migrate applications from one server/platform to another and servers form one network segment to another? For applications that can't be re-IP'd (many exist in this category), the general answer is to nuke and pave (or extend a clusterable application, then remove the segment that needs to be "moved"). For "normal" applications (httpd, mail, directory services, etc), what are the checks ou perform before, during, and after a move to ensure the health of the migrated app/server? An example with Apache: backup httpd conf directory change httpd conf files to use new IP address of server change (or add) IP of server restart Apache verify web server still serves pages reboot server verify environment comes back up healthy

    Read the article

  • What are the challenges when my enterprise desires to move the processing component of an applicatio

    - by Berkay
    Assume that i have an enterprise accounting application that consists of a front-end interface, a processing tier, and a back-end database. This is an application that contains private business data, and thus is traditionally run in a secure private network environment within the enterprise. What are the challenges that appear when my enterprise desires to move the processing component of this application to a cloud computing data center in order to achieve greater scalability or to reduce IT costs ? Pls note: do i have to make significant changes to my own infrastructure to enable external access to formerly private resources? do i have to modify the application code to handle new network topology ? thanks, if you give your answers in a simple manner, really appreciated.

    Read the article

< Previous Page | 341 342 343 344 345 346 347 348 349 350 351 352  | Next Page >