Search Results

Search found 14975 results on 599 pages for 'os x'.

Page 434/599 | < Previous Page | 430 431 432 433 434 435 436 437 438 439 440 441  | Next Page >

  • What is the best way to get the external internet gateway IP reported periodically?

    - by basilmir
    I have a OS X Server behind an airport extreme, serving services via opened ports on the airport. The server has a 10.0.x.x local address, always the same one. The airport extreme gets it's external IP address via PPPoE, and sometimes... once a week it changes. For security reasons WE ACTUALLY like this behavior. But i need a way to know the external IP address just in case i need to connect and do something to the server while on the outside. What can i do?

    Read the article

  • Hard drive placement

    - by zm15
    I'm a video editor working with large HD files. I am building a new computer and need some help. I will be running 2 hard drives. One with the operating system and all the programs. And one with all the project files I will be working from. I am keeping these seperate. I will be purchasing a 10k rpm hard drive. So i will have a 10k rpm drive, and a 7200rpm drive. Should I put the OS on the faster drive, or put my working files on the faster drive?

    Read the article

  • Can't resolve Mac's machine name on VPN

    - by Raghuveer
    My mac'c machine name is something like this: hostname.company.com but whenever I connect to VPN, it becomes something like vpn-xxxx.company.com where xxxx are some random numbers. Because of this, some of my scripts which are dependent on host name gets blocked. We use the standard mac's vpn setup which comes with OS X Lion (under network preferences). How can I resolve to the correct mac's name even if I am on vpn ? That is even if I am connected to VPN, my machine name should resolve to hostname.company.com and NOT to vpn-xxxx.company.com. Any suggestions would be really appreciated.

    Read the article

  • Storage replication/mirror over WAN

    - by galitz
    Hello, We are looking at storage replication between two data centers (600km apart) to support an active-passive cluster design for disaster recovery. The OS layer will be mostly Windows Server 2003/2008 with some OpenSuSE Linux used for performance monitoring on VMWare or possibly XenServer. The primary application service to replicate is Nvision. Datacenter 1 will have two storage systems for local active-passive or perhaps active-active replication with Datacenter 2 used as a last resport disaster recovery site. We have a handle on most aspects, but I am looking for specific recommendations on storage platforms that can handle remote replication cleanly. Thanks.

    Read the article

  • Automatic switching between monitor configurations

    - by Michael Aquilina
    I have a laptop and an external monitor and i was wondering if there was a simple approach to switching between multiple monitor configurations based on the detected available displays. For example: When i am at home and i plug in my external monitor i would like this to automatically become enabled and the laptop screen to become disabled. As soon as i pull out the display cable for the external monitor, i would like the laptop screen to automatically become enabled. I was expecting this to just "work" just like it does in windows - but it seems to be much harder than that. I am aware of the xrandr command to turn displays on and off but i cannot seem to find a way to get this to work the way i describe above. I had also found this post about switching between multiple monitor configurations and the results seem a bit inconclusive. However i was hoping that with xrandr there would be a simpler solution. For me, the fact that when i pull out my external monitor the screen just goes black and i get an error message is a big issue holding me back from making the complete switch to linux as i move around alot as a student. My OS of choice is currently Kubuntu 12.04 but i am willing to change to something else if it provides a better way of setting up the described setup.

    Read the article

  • How can I prevent JungleDisk/MacOS X (10.6) creating a local volume for a removed external drive?

    - by Rew
    Ok, here is situation: I use JungleDisk to sync an online folder on to a external drive connected to my Mac. If I right click Finder, click Go to Folder... then type /Volumes/ I see the drive linked here. Once I remove the external drive, an actual folder is created here in the name of the external drive, JungleDisk continues to copy files to this folder, rather than stop. Is this a feature of Mac OS X? Can I turn if off? After I re-connect my external drive, the link to the drive is appended with a 1 (so if I called the drive SpareDrive it becomes SpareDrive 1 as the newly created folder is called SpareDrive. I realise my explanation isn't very clear, but anyone understand this, and knows how to prevent it happening please let me know. PS: I have a low reputation as I don't use this often, I tend to use stackoverflow, but will check back here for answers.

    Read the article

  • How much CPU use is too much?

    - by Jonathan Sampson
    I've got a server that receives around a million unique visitors a month and I've recently began using Plesk to help monitor some of the vitals on the box itself. RAM I can make sense of, but I'm not really sure if my CPU usage is too high, low, or about average for this number of visitors. The server exists solely to serve up a somewhat hefty WordPress blog. This is one week. What types of things should I look out for? Some other information about this server follows: VCPU(s): 4, RAM: 6GB, HDD: 30GB, OS: Ubuntu Server 10.04 x86_64

    Read the article

  • Windows 7 – Fun with VHD

    - by guybarrette
    I’m teaching about TFS 2008 next week and I wanted to use TFS in a virtualized environment so I downloaded the TFS + Team Suite VPC image from Microsoft’s Website.  Working with Windows 7, I opened the VM with the built-in Windows Virtual PC.  The VM loads fine but the problems started when I tried to install the VM additions: I simply couldn’t get them to install properly. I then looked at VMware and found that they have a product called VMware Player that can load Virtual PC VMs.  Tried that but VMware Player failed in converting the VHD. I then looked at VirtualBox.  Created a new VM, attached the VHD and bingo!  Worked like a charm.  The only real caveat is that the guest Windows will ask for the OS CDs to install new drivers so you must have either the CD/DVD or the ISO file (sweet!) to proceed. OK, I got it working in VirtualBox but I’m curious why I couldn’t install the additions from Windows 7 Virtual PC onto a Windows Server 2003 VM.  Anyone has a clue? BTW, thanks to Rolly Perreaux who pointed my to his blog where he goes into great details explaining how to use VM images with VirtualBox.  Good stuff! var addthis_pub="guybarrette";

    Read the article

  • How to stop LDAP authentication in ubuntu?

    - by Kery
    My OS is Ubuntu 12.04 and use LDAP authentication. Now I meet a problem that another people want to access my system. But he is in another domain so he can't login. And I have no right to change this configuration in LDAP server. So I have to choose a workaround to solve this problem, for example close the LDAP authentication and use local authentication (I have root right in my system) or create another account which is not registered in LDAP server (I did this but can't change the created account password. The error is 'password reset by root is not supported'). Of course any other suggestion is appreciated! Than you in advance!

    Read the article

  • Mirroring of Apps across servers

    - by user1038814
    We wish to host multiple apps across multiple servers. What we are looking for (ideally) is an existing solution which will work. For example, normally to do it we'd follow a route (for failover) like: App is installed on one server along with mysql database App is also installed on a second server. Rsync is used to mirror the files over to the second server and ensure consistency MySQL is installed with a Master-Slave setup. We use a service such as DNS Made Easy which has a DNS failover. If one server goes down it automatically routes traffic to the backup server We have done the above a few times and generally its fine. The issue I have here is that the above is for one app. What I would like to look at is how we can manage for multiple apps and if there is a layer (such as VMWare) that has complete mirroring built in at the OS level? For example how do web hosts currently do it when they ensure that more than one machine is running a bunch of hosted websites. If you were running hosting and you had 200 clients on a server you would want the same clients across 2 or more servers and want everything mirrored. Any advice would be much appreciated.

    Read the article

  • Shut down by pressing the power button twice

    - by iHarry
    I have set my laptop's power button to shut down the computer when it is pressed (Windows 7 x86). But I often hit the button accidentally, and it's a rare case that I'm able to actually prevent it from shutting down. I don't want to disable it though. Is there a way so that I get a confirmation whether I really want to shut down the computer (just like in Mac OS X Lion, where a confirmation appears when the power button is pressed)?. Is there some way to bring the same to Windows 7 (and 8)?

    Read the article

  • Switching to Chrome from IE

    - by Alan Parrish
    Hi, I work in IT at a school and we recently updated our database software, however the web access that the teachers use to do registration does not work too well with Internet Explorer 7 and we're thinking about switching to chrome (mostly due to me hating IE in general) but the problem we have is that users are unable to migrate their bookmarks over from IE due to account restrictions, is there any way to get this working? A bit of info about the system if it could be helpful; Almost all the client PCs are running Win XP SP2. My colleagues machines are running Windows 7, I use OS X Snow Leopard and most of out servers are running windows server 2003 (with the exception of 2 one on 2000 server and another on 2008 R2). The Active Directory Controller is running Windows Server 2003

    Read the article

  • Best practice for administering a (hadoop) cluster

    - by Alex
    Dear all, I've recently been playing with Hadoop. I have a six node cluster up and running - with HDFS, and having run a number of MapRed jobs. So far, so good. However I'm now looking to do this more systematically and with a larger number of nodes. Our base system is Ubuntu and the current setup has been administered using apt (to install the correct java runtime) and ssh/scp (to propagate out the various conf files). This is clearly not scalable over time. Does anyone have any experience of good systems for administering (possibly slightly heterogenous: different disk sizes, different numbers of cpus on each node) hadoop clusters automagically? I would consider diskless boot - but imagine that with a large cluster, getting the cluster up and running might be bottle-necked on the machine serving the OS. Or some form of distributed debian apt to keep the machines native environment synchronised? And how do people successfully manage the conf files over a number of (potentially heterogenous) machines? Thanks very much in advance, Alex

    Read the article

  • Is there a way to schedule an automatic WinClone run on my Bootcamp partition?

    - by user17873
    Hi, I've currently got time machine setup to back up my entire OS X installation. I also have a backup tool within my Bootcamp Windows 7 installation which automatically backs up my windows profile data to an external drive partition. Finally, I'm also backing up my Bootcamp partition weekly and storing on an external drive using WinClone. The final piece I need to complete my external backup process is to have the WinClone application backup my Bootcamp partition automatically once a week rather than having to call it manually and remember. Is this possible?

    Read the article

  • cookie not being sent when requesting JS

    - by Mala
    I host a webservice, and provide my members with a Javascript bookmarklet, which loads a JS sript from my server. However, clients must be logged in, in order to receive the JS script. This works for almost everybody. However, some users on setups (i.e. browser/OS) that are known to work for other people have the following problem: when they request the script via the javascript bookmarklet from my server, their cookie from my server does not get included with the request, and as such they are always "not authenticated". I'm making the request in the following way: var myScript = eltCreate('script'); myScript.setAttribute('src','http://myserver.com/script'); document.body.appendChild(myScript); In a fit of confused desperation, I changed the script page to simply output "My cookie has [x] elements" where [x] is count($_COOKIE). If this extremely small subset of users requests the script via the normal method, the message reads "My cookie has 0 elements". When they access the URL directly in their browser, the message reads "My cookie has 7 elements". What on earth could be going on?!

    Read the article

  • Amazon - install a complete server on EBS

    - by user1169575
    is it possible to install a full working OS with a webserver, db, and all needed stuff on an EBS storage? If so, would I immediatly gain benefits by mounting this EBS on a better instance? Otherwise (if I cannot install a complete image, or if you don't think it's reasonable) can I install the software so that I only need to mount the EBS on a new instance to have it working? I purchased a Medium Reserved Instance, but when there will be the need to get a better instance I'd like to move the whole db/website, I'd simply buy a better instance and then attach the EBS. Is it possibile? I'm imaging about it like an hard drive that would be mounted on a better server. Of course, more RAM would allow me to increase caching limits, and that's ok, but I don't want to reinstall anything (the main website is a magentocommerce and it's pretty painful to move it). P.S. is the Standard EBS (100 IOPS) valid or do I need to choose a Provisioned IOPS (up to 1000 IOPS)?

    Read the article

  • VPC on Windows 7 very slow network

    - by Shigg
    I have a Windows 2003 virtual machine which I use for website testing. I've just installed Windows 7 and am using the new version VPC (not xp mode). When I try to copy a file - I need to copy some big databases across - I get a file copy speed of about 20k per sec. Copying from one PC to another on the real network transfers files at 13mb per second. Any ideas what may be causing this? I've turned off differential network compression on win 7. The Virtual HD is on a seperate physical drive to the OS. Running Windows 7 64 bit on a dual xeon with 16gb ram and 10,000 rpm drives. Tried installing VPC 2007 but windows blocks it running saying its not compatable. Many thanks for any ideas.

    Read the article

  • All browsers hang on windows 7 in tab says 'Waiting for...' need restart

    - by Lakshmila
    I'm having a frustrating problem with browsers on OS windows 7 home premium. IE 10: My web browsing works ok for a while and then pages freeze and tab at top says "Waiting for ...' and never loads. Strangely I can still get to Google search results but if I click anything it freezes. Chrome freezes too and says 'The page has become unresponsive ...kill page' Safari stops also. I have tried clearing cookies, cache, turning off Bitdefender firewall and nothing apart from a restart fixes it. consequently I am forced to restart my computer several times a day to stop this browser hang. I have no idea what could be causing this ..any help greatly appreciated.

    Read the article

  • Change Apache DocumentDirectory path in trueCrypt partition

    - by Alan C
    Hello, I'm recently moving from windows to linux, so I've setup my machine to dual boot Windows7 and Ubuntu 10.04. I was able to successfully setup Apache on the Ubuntu partition, but I need to move the DocumentRoot since my websites are on a TrueCrypt partition that is in another hard drive so I can have them accessible in both OS. I followed some guides on how to change the path for the DocumentRoot so I end up modifiying the default file at /etc/apache2/sites-available DocumentRoot /media/truecrypt1/www <Directory /> Options FollowSymLinks AllowOverride None </Directory> <Directory /media/truecrypt1/www/> Options Indexes FollowSymLinks MultiViews AllowOverride None Order allow,deny allow from all </Directory> Those are the lines that I've changed, but now when I go to localhost I always get the Forbidden You don't have permission to access / on this server. Apache/2.2.14 (Ubuntu) Server at localhost Port 80

    Read the article

  • SQL Server 2005/2008 Licensing Decision

    - by Hakim
    Hello, I have purchased a dedicated server from a reputable Hosting company. They only have Windows Server 2008 OS installed on it and NO Sql server. Server Configuration is Intel Dual core Processor with 2GB of RAM and 100GB HDD. I wanted to host my web services on that server which will be using the MS SQL Server 2005 at the backend.There are multiple web services and each using a different Database. Microsoft has CAL basis Licensing , Which I understand is based on number of users accessing the database directly ( I may be wrong ) . But my users will be accessing the webservice and no direct connection to the database as such. Yes but the number of users accessing the web server cannot be known and is not under my control. Which Licensing is best suited for this kind of setup ? I don't need analysing and BI services right now ,but i may want to upgrade that in future may be. Any help will be appreciated. Thanks

    Read the article

  • HTTP PHP Authentication and Android

    - by edc598
    I am working on a website for which I hope to have an application for as well. Because of this, I am creating PHP API's which will go into my Database and serve specific data based on the method/function called. I want to protect these API's from misuse however, and I plan on implementing Authentication Digest to do so. However one of the OS's I want to support is Android. And I know that a malicious user would be able to reverse engineer the Android app and figure out my authentication scheme. I am left wondering: 1. Is there a better way to protect these API's from misuse? 2. Is there a way to prevent a malicious user from reverse engineering the app and potentially seeing the source code for it, enabling them to see my authentication scheme? 3. If none of these are preventable, then is my only option to have a Username/Password cred specifically for the Android app, and when eventually hacked, change the creds and issue an update for the app? I apologize if this is not the place to post such a question. Still pretty new to StackOverflow. Thanks in advance for any insight, it would be quite helpful.

    Read the article

  • How do I get Acrobat Pro X to close files without quitting?

    - by CyberSkull
    Adobe Acrobat Pro X doesn't seem to want to relinquish files after I close them. When working on some PDFs after I save and close them, I then go to move (not copy) them to the appropriate done folder on our server. Then I get a series of errors saying that Windows cannot move the file as Acrobat Pro X has it open. Looking at the Acrobat window, the file is not loaded. Is there a way to get Acrobat Pro X to close the file like normal (without quitting the program) and actually release the file to the OS?

    Read the article

  • determine the archetecture of a mac from the command line or script?

    - by Brian Postow
    I'm writing a shell script, and I need to know the archetecture, ie PPC or Intel. Back in the day, there was a program /bin/arch that told you, but my mac doesn't seem to have it... Is there an easy way I can do this? Grep for something in a logfile? call some other program that spits that out as a side effect? It would be nice to know what OS Version I'm running too, but that may not be necessary. thanks

    Read the article

  • Visual Studio Development on Virtual Box, Boot Camp, or VMWare Fusion

    - by Eli
    I currently have a Mac, 2ghz and 2 gigs of ram, running OS X Leopard and Virtual Box with a Windows 7 Pro 32bit virtual machine. Performance on the virtual machine is fine for minor tasks but is very clunky while trying to multi-task or develop in Visual Studio 2008. What would be my best option for being able to use Visual Studio, keeping cost and time in mind? 1) Upgrade ram to 4 gigs ($100). Will this really improve my performance enough to use Visual Studio in a Windows 7 vm? Or am I just wasting time/money? 2) Reinstall/restore Windows 7 disk image as a Boot Camp partition. I assume this should improve my performance, yes? 3) Purchase VMWare fusion instead of VirtualBox. Does Fusion require less resources to run? I am open to any suggestions. Thanks in advance

    Read the article

  • XAMPP on windows 7 not working

    - by 404Error
    Hey there, I just installed XAMPP lite on Windows 7. I have two drives - C: for the OS and regular files, and an external drive E:. I insatlled XAMPP lite on E: (on the root), and its been giving me problems. Apache works well enough, but MySQL doesn't work. When I go to http://localhost/phpmyadmin/, it gives me the following error: Error MySQL said: #2003 - Can't connect to MySQL server on 'localhost' (10061) Connection for controluser as defined in your configuration failed. Any ideas as to what could be the problem? I used the zip file for XAMPP lite, the 32 bit version. This is on Windows 7 Home premium. Thanks!

    Read the article

< Previous Page | 430 431 432 433 434 435 436 437 438 439 440 441  | Next Page >