Search Results

Search found 42646 results on 1706 pages for 'vbox question'.

Page 806/1706 | < Previous Page | 802 803 804 805 806 807 808 809 810 811 812 813  | Next Page >

  • Not able to login to new AMI on EC2 - moveing from micro to small instance

    - by zengr
    I had a t1.micro instance (old_server) on Linux on EC2 and now I need to upgrade my server to m1.small Linux (new_server). So, here is what I did: Shutdown old_server and create an AMI. Launch the new AMI with m1.small configuration. (I kept the key and security group same as old_server) I tried to login by: ssh -i my_key.pem [email protected] But, it gives a connection timeout error. My login to old_server works fine. So, my question is: What is the correct way to scale up (vertically) an EC2 instance? Where am I going wrong in the above mentioned step? When I create an AMI in step 1, is the EBS (data) also copied.

    Read the article

  • Restrict Overriding memory_limit in php.ini

    - by Rohit Batra
    We have a issue that just came to our notice, some of our developers used ini_set('memory_limit', -1) to over-ride the php.ini memory_limit settings in their code files. Due to which sometimes our Memory/CPU usage is way far more than usual and we even have experienced server hanging/crashing many times. Is there any way to restrict these changes for certain params like memory_limit, upload_max_filesize & post_max_size? Users should not be able to edit them after a certain limit set? Sorry if i asked a dumb question, tried searching but there was no reference to this only found answers asking for increasing these parameters.

    Read the article

  • Can / how to install openSUSE as a GUI-less home server with only wifi?

    - by Dougal
    Hello there, I'm following an article to set up openSUSE as a commandline server (http://www.howtoforge.com/perfect-server-opensuse-11.1). It seems to only work for wired internet connections. Is there any way to install openSUSE as a commandline server in my home network but using a wifi internet connection? Preferably from the get-go and not simply installing with LAN cable internet and then installing some wifi things later. Or, perhaps, to rephrase the question "How can I get the openSUSE install / post-install procedure to recognize my wifi internet connection?" Kind regards, Dougal

    Read the article

  • How to decide where to purchase a wildcard SSL certificate?

    - by user664833
    Recently I needed to purchase a wildcard SSL certificate (because I need to secure a number of subdomains), and when I first searched for where to buy one I was overwhelmed with the number of choices, marketing claims, and price range. I created a list to help me see passed the marketing gimmicks that the greater majority of the Certificate Authorities (CAs) plaster all over their sites. In the end my personal conclusion is that pretty much the only things that matter are the price and the pleasantness of the CA's website. Question: Besides price and a nice website, is there anything worthy of my consideration in deciding where to purchase a wildcard SSL certificate?

    Read the article

  • Is it a good idea to have the operating system on a solid state drive?

    - by Kenji Kina
    There is something I don't quite understand. I know a SSD helps with OS load times, but I'm not sure if all this boost is only noticeable/interesting when booting, or gives an all around considerably better experience thereafter. I am interested in having a quick and responsive environment after booting, which leads me to think that it'd be better to spend the SSD capacity in my most used apps (and the page file? Another inside question) and not the OS itself. This, of course, means that I don't know just how much the OS reads/writes its files during normal usage. So, how good an idea is it to dump the whole 20GB+ of Windows 7 OS into the SSD (considering the hefty price per GB of SSD capacity) if I can put up with the usual hard disk boot times? Would I be missing on a lot if I didn't?

    Read the article

  • reg delete gives me "access is denied" but regedit delete is ok

    - by Radek
    I need to delete a key from a command line. So I wanted to use reg delete "the key to be deleted" /f but I get ERROR: access is denied. From the same login session (the same user) I am able to delete the key without any troubles from regedit.exe that is not run as administrator. I cannot use runas command to execute reg that I believe would be to solution because in fact I want to use reg to delete registry entry for administrator profile so runas works again. More info in my other question Windows7 corrupted profile - prevention exists?

    Read the article

  • Firewall Deep Inspection Updates and Antivirus Subscription, worth it?

    - by msemack
    I realize that this is a subjective question, but I'm trying to get some experiences We have Juniper firewalls in our organization (SSG-320M, SSG-5, and some old NS-5GT). We have the option of a yearly subscription for: Deep Inspection Signature Updates Juniper-Kaspersky Antivirus I seem similar services available from other Firewall vendors. We have Symantec Endpoint Protection deployed to all workstations and servers, plus a dedicated appliance for e-mail spam/virus filtering. So, I'm not sure what these firewall-base services will bring to the table that I don't already have. I would appreciate some feedback from people using these firewall services (Juniper or otherwise). Are these services generally worth it? Do they really catch anything? Do they interfere with normal traffic (false positives)?

    Read the article

  • How to quickly switch monitors on multi-monitor setup? (W7)

    - by Saxtus
    At my desktop PC, I have 3 displays hooked up on my Palit GeForce GTX 480 to extend my desktop: DVI: Samsung SyncMaster T220 (22" 16:10, 1680x1050, as my main) DVI: Samsung SyncMaster 940BF (19" 4:3, 1280x1024) HDMI: Sony Bravia KDL-32D3000 (32" 16:9, 1360x768) As expected, the GFX card can only utilize 2 of the displays at any given time, so I have to manually switch using Windows' "Display Properties" dialog, each time I want to switch between them. I've tried UltraMon, DisplayFusion and PowerStrip (so please don't refer me to a question where those are given as solutions) but they are unable to switch monitors. Instead they can only change display resolutions and settings on the existing active monitors, with specially weird results when I've saved the monitor states and then tried to switch monitors using the saved states! Any idea on how to bypass the awful "Display Properties" dialog and it's "confirm your change" timer? Thanks in advance.

    Read the article

  • How do I troubleshoot CanoScan LIDE20 on Ubuntu 9.10 in VirtualBox?

    - by kara-marfia
    This may be a VirtualBox question more than anything else. The host OS is Win7, which sees the scanner, but has no drivers for it. VirtualBox sees & recognizes the scanner as "Cannon Scan" and shows "state captured" when I hover over USB devices. Searching gives me many happy reports of this particular scanner being well supported, and that SANE should work with it immediately (also listedin the Supported Hardware List. I get "no devices available" from XSane... and I'm not sure where else I can check. I'd like to see if VirtualBox is telling the truth and find out if the USB port is actually being fed into the virtual OS. Can anyone point me in the right direction of where/how to do that in Ubuntu? Or other bright ideas. I'm starting with a little project to add some zing to my beginner linux studies - I'm sure I haven't checked some obvious things, because I don't know them yet. ;)

    Read the article

  • SSL Certificate Stops Working after Server Reboot on IIS7, W2K8

    - by Zac
    We recently upgraded from W2K3/IIS6 to W2K8/IIS7 and have been having problems with our SSL Certificate (Thawte 123 SSL certificate) ceasing to work after rebooting. Initially, the intermediate certificates would stop working and we could repair the problem by reinstalling all of them after the reboot (annoying, but not the end of the world). Unfortunately, this is no longer working. The certificate chain has been doublechecked by several tools and people with decent knowledge but no one has been able to identify the cause of the problem. The bindings in IIS have been checked as well The cert itself is also still valid. NOTE 1: I have seen THIS question which seems to be very similar, but there is no satisfactory answer in that post and it's a year old so not likely to get one any time soon. NOTE 2: I'm asking this on behalf of a co-worker so won't be able to provide instant feedback to any questions/suggestions but I will pass it on. The url is: http://www.flirtalike.com / https://www.flirtalike.com Screenshots:

    Read the article

  • network user isolation

    - by seaquest
    My question is for a network with a Linux iptables router gateway. How can it be possible to prevent inter-network traffic of those users. Think this case as a public network, IPs are distributed through linux gw and users are authenticated thru the gateway. We want to protect public users from public users. Network is not wireless and I can not use Wireless AP user isolation. Actually I have a simple method. Subnet the network into /30 mask. Give minimum IP of each subnet to the gateay and ditribute those /30 IPs from the subnet. But this is pretty costly for such an aim. I want to ask for other methods Thanks.

    Read the article

  • How to set up a staging apt repository to securely manage upgrades

    - by andreash
    Hello, I would like to be able to run automatic apt-get upgrade (once per hour) on our servers (Ubuntu 10.04), so that I don't have to do it manually on all of them (about 15). However, for production machines, that's not a good idea ... So here's my idea: Set up a local repository for all 'approved' updates for critical packages. I would then push updated packages from upstream to our local repo after I tested them, and all servers could automatically (apt-cron?) upgrade from this repository. So my question is this: How do I configure apt on the clients so that they use the local repository only for all packages which exist on the local repository, and the upstream one for all other packages? Does this actually make sense? Or am I missing something? Anyways, thanks for your insight! Andreas.

    Read the article

  • dual/multi-boot computers and software licensing

    - by Matt
    Suppose you have a computer with two or more operating systems, and a certain piece of software whose license terms allows it to be installed on one computer, and it does a daily check with a remote server to verify that your serial is only used on the original install computer. You install this software on each of your OSes, but since its a different OS the remote server would have to determine that it is not on the same computer, and so would disable your license. So my question, when a license refers to a single computer, does a situation like this usually count as a single computer, or do the multiple OSes sort of make it multiple computers? How do you think a software vendor (specifically thinking AV companies that do this sort of serial check) would handle this situation?

    Read the article

  • Samba Server as domain controller

    - by garden air
    I am using centos 6.2 on my system. I want to make samba server as PDC to join the clients computers operating systems i.e xp,windows 7 and share their files & directories.Currently I have 200 PC running both windows xp and win 7. The question I want to as is does samba 3.5.10 has a capacity to join 200 computers as a domain controller & authenticate the users ? thanks garden Thanks for your guidence.Well at the moment I am using CentOS 6.2 and samba is installed using yum command. It is amazing that I can join samba as a domain comtroller of 200 clients machines. By the way what is samba maximum limit for joining number of clients PC's.

    Read the article

  • CentOS Vs Windows Server 2008

    - by Steve
    Hi, Apologies if the question appears ambiguous, I have little experience in this area and was after some informed opinions. I am deploying a test scenario of a server/client network and need to make some choices for Server. The client will be a Windows system as it meets the requirements for the client, the server choice has more room for selection. From my experience with Linux in general and the appealing nature of open source for low cost, security etc and the availability and performance of database and web server programs I have been considering CentOS as a server choice. I have the ability to make most of the choices of what software / server packages I wish to install. This includes Active Directory (something I have no experience with). How well does this operate with Windows clients? Am I being too selective and creating unnecessary complication by setting out not to use a Windows Server OS?

    Read the article

  • Distribute outgoing connections among multiple IPs configured on the same NIC

    - by cedivad
    I have a NIC with 2 aliases on it. The network interface has 3 IPs configured on it. Think about it like this: i can ping the same server by hitting .100 .101 and .102. I want the source address of the outgoing connections to be distributed among these ip. So if i have 3 opened connections, one connection will have result as having an IP address ending with .100, the other two should result as having as ip addresses .101 and .102. I'm using FreeBSD but I think this question to be Linux-Like wide.

    Read the article

  • Best practices for setting lm-factor in Squid refresh patterns

    - by Mpentecost
    I am running a Squid (3.1) cache in front of Django. The content of the site does not change very often, so Squid gives our backend much needed breathing room. Currently, this is the refresh pattern that we are using to cache the content: refresh_pattern . 60 100% 60 We basically want to cache everything for at least an hour (and only an hour) before Squid then re-validates the content. My question is on the "100%" parameter, which sets the lm-factor. I'm not sure if setting that to 100% is doing what we want it to. The assumption was that by setting it to 100%, it would ensure that objects stay in the cache for the max cache time. Is this an incorrect assumption? What are the best practices that one should follow when setting up a refresh pattern like this?

    Read the article

  • Optimum number of threads while multitasking

    - by Gun Deniz
    I know similar questions have been asked but I think my case is a little bit diffrent. Let's say I have a computer with 8 cores and infinite memory with a Linux OS. I have a calculation software called Gaussian that can take advantage of multithreading. So I set its thread count to 8 for a single calculation for maximum speed. However I really can't decide what to do when I need to do run for instance 8 calculations simultaneously. In that case should I set the thread count to 1(total 8 threads spawned in 8 processes) or keep it 8(total 64 threads spawned in 8 processes) for each job? Does it really matter much? A related question is does the OS automatically does the core-parking to diffrent cores for each thread?

    Read the article

  • Can I block social share buttons with Privoxy?

    - by gojira
    A great many pages have these "Like", "Tweet", "G+1", "share" row of buttons all over the place and in each post in threads. Can I block these unwanted context with Privoxy? I am already using Privoxy and it blocks a lot of unwanted content, but still these "social" buttons are all over the place. I want to completely remove these buttons specifically by using Privoxy. I know that it is possible to block using AdBlock LITE and other software, but my question is specific to Privoxy (reason, I want one point to block all unwanted content and it needs to work on devices / softwares which do not have AdBlock LITE, therefore I use Privoxy). -- Software used: Privoxy 3.0.21 under Windows XP

    Read the article

  • Windows replacement for HAProxy

    - by GrayWizardx
    It does not appear that there is a similar question to this already posted, so I will go ahead and ask. I am working on a project that could benefit from having two - four servers handling incoming requests to a backend webservice. The service does not require SSL but does need to support occassional long running processes (upto 120 secs). This project does not at present have the funding to purchase a hardware load balancing solution. I have previously used HAProxy as a solution for this, and found it very simple and straightforward. Is there a similar product for windows (server 2003 or 2008) which provides similar configuration options and runs as a lightweight service? For reasons outside my control I cannot setup a Linux machine (physical or virtual) and so I am looking for behaviour that can be deployed on a windows machine. I can only find Perlbal which appears to fall into this category. So as not to keep this open indifenitely I will give credit to the only answer.

    Read the article

  • Move samba and ldap config to new machine

    - by Mateusz W
    I'm not advanced administrator so this may be n00b question but I'll be grateful for help. I have following problem: I have to move my server (ubuntu server) from old machine to new. I have samba & ldap on old server, so I've copied the configuration files from the old machine to new machine, made smbldap-populate, and when I use smbldap-userlist on the new machine it lists all users from the old machine. But when I try to log in as any of these users, or show its id (id username) I get no such user error. What more I have to do to use these accounts on the new machnie?

    Read the article

  • *.example.com wildcard domain can be parsed from a single page?

    - by Sean Kean
    For a domain 'example.com' - what is the easiest way to set up a wildcard dns (*.example.com), hosting, and htaccess/httpd.conf/virtualhost, and script on a page so that: how.do.i.setup.a.site.with.wildcards.like.this.example.com or anything.that.is.given.as.a.subdomain.for.example.com is rendered by a page at example.com/index.html - yet keeps the wildcard subdomain in the URL bar and passes the full URL as a parameter for rendering tags in HTML? An example tag is a Facebook comment: { div class="fb-comments" data-href="http://how.do.i.setup.a.site.with.wildcards.like.this.example.com" data-num-posts="2" data-width="500" } I just opened a hosting account with spry.com and have a VPS running Ubuntu 11.04-x86-LAMP - Essentially, what is the most straightforward way of doing this? Thanks so much. (I originally posted this over on stackoverflow but realize its more of a serverfault question)

    Read the article

  • Download a file via HTTP from a script in Windows

    - by Jason R. Coombs
    I want a way to download a file via HTTP given its URL (similar to how wget works). I've seen the answers to this question, but I have two changes to the requirements: I'd like it to run on Windows 7 or later (though if it works on Windows XP, that's a bonus). I need to be able to do this on a stock machine with nothing but the script, which should be text that could be easily entered on a keyboard or copy/pasted. The shorter, the better. So, essentially, I'd like a .cmd (batch) script, VBScript, or Powershell script that can accomplish the download. It could use COM or invoke IE, but it needs to run without any input, and should behave well when invoked without a display (such as through a telnet session).

    Read the article

  • When connected through a router, a new gateway appears, and I lose my internet connection.

    - by idan315
    I have two computers connected to a router, connected to my ADSL modem by PPPoE. One of those computrs use Windows XP, the other uses Windows 7. The computer using Windows 7 keeps getting limited or no connectivity, and I can't use it to connect to the internet. Connecting the computer to the modem without the router 'in the way' worked, and I was able to use the internet - which led me to believe that the problem is with my router. However, I have changed my router, and the problem persists. (Just to be clear: the computer using Windows XP can connect to the internet without any problem). This entire question is about the PC using Windows 7. The only symptom of this problem is that when I see my network devices, in addition to my router (which appears as "Network 4", or something of the sort), I also see an "Unidentified Network". When using ipconfig to check the problem, I see a second gateway, with the IP 0.0.0.0, in addition to the IP of my router.

    Read the article

  • Why Standards Place Limits on Data Transfer Rates?

    - by Mehrdad
    This is a rather general question about hardware and standards in general: Why do they place limits on data transfer rates, and disallow manufacturers from exceeding those rates? (E.g. 100 Mbit/s for Wireless G, 150 Mbit/s for Wireless N, ...) Why not allow for some sort of handshake protocol, whereby the devices being connected agree upon the maximum throughput that they each support, and use that value instead? Why does there have to be a hard-coded limit, which would require a new standard for every single improvement to a data rate?

    Read the article

< Previous Page | 802 803 804 805 806 807 808 809 810 811 812 813  | Next Page >