Search Results

Search found 11930 results on 478 pages for 'shared machines'.

Page 245/478 | < Previous Page | 241 242 243 244 245 246 247 248 249 250 251 252  | Next Page >

  • CentOS: revert python version back to original

    - by NP
    Hi all, I installed python 2.6 using the instructions here on CentOS 5.4. However I realized it was a bad move and I need to revert back to 2.4, which was there originally. Can anyone guide me on how to undo what I did here? In particular, I am not sure how to undo this: Configure ld to find your shared libs: $ cat /etc/ld.so.conf.d/opt-python2.5.conf /opt/python2.5/lib (hit enter) (hit ctrl-d to return to shell) $ ldconfig I tried removing the alias and the symlink and even re-aliasing python to /usr/bin/python, but when I try to install an RPM i get this error: error: Failed dependencies: libpython2.4.so.1.0 is needed by ... Thanks in advance.

    Read the article

  • Slow git clone and fetch

    - by EtienneT
    I setuped gitosis on a linux server following this tutorial: http://scie.nti.st/2007/11/14/hosting-git-repositories-the-easy-and-secure-way We are using git on our windows machines with TortoiseGit and msysgit. Pushing changes to the server is pretty fast, but when we want to clone or fetch changes from the remote server, it begins really fast (800k/s) and then drop pretty fast to around 3 to 30k/s and it can take forever to update. git-pull for small update is fast, but as soon as we have to download something of more than a few MB, it is slow. We are switching from SVN to git and this is holding us back from using git full time. Thanks!

    Read the article

  • DNS responding only to nslookup in Windows 7

    - by DeeJay1
    Hello. I have an interesting DNS issue in AD (2008R2) which I can't find a way to debug. Mainly if from commandline I try ping <machinename> then I get the reply that the hostr isn't found. But when I do nslookup <machinename> I get the correct response from my DNS with proper AAAA and A records. Anyone has an idea what could cause this? The client machines run Windows 7 Enterprise.

    Read the article

  • PHP-APC is exceeding limit apc.shm_size often

    - by user142741
    I am implementing the apc on a shared server that currently has 1000 sites (using wordpress, moodle, etc.). I'm Looking for the admin page, and I see "Cache full count" is growing rapidly. I've tried increasing the value of "apc.shm_size" reduce value "apc.ttl" increase the value "apc.shm_segments" but I can not resolve this issue. What am I doing wrong? I'm putting down some information: apc.ini: extension=apc.so apc.shm_size = 256 apc.enabled = 1 apc.ttl = 300 apc.user_ttl = 300 Ubuntu: 12.04 PHP: 5.3.10 APC: 3.1.7 Server has 16GB memory Limit share memory: 256MB

    Read the article

  • Cannot access Windows 7 share from Windows XP

    - by artfulrobot
    I have a new Windows 7 machine named PAP44 in the PAP workgroup. The networking is set to "Work" mode for the wired LAN. I have a couple of users and I've shared a folder and set it so both users can read/write. Confusingly for me, rather than sharing just that folder (as I'm used to with older versions of Windows) it appears to be sharing a path (\\pap44\users\...\myFolder) From another machine on the LAN, running XP, when I go to \\PAP44\Users I'm asked for a username and password, but neither of the usernames+passwords work. It just jumps back to the username and password dialogue, except that the username I entered gets prefixed with PAP44\ My end goal is to get my Debian/Ubuntu machines to be able to access this share, but first of all I thought I'd try to get it working in Windows, after all, that's supposed to be easy! Is there another step? (PS. I am not a "hit and run" case!)

    Read the article

  • I want to replace Windows Server by Samba 4 for authentication purpose [closed]

    - by garden air
    I am using windows Server 2008 in a mediam size organization. I have 200 clients at the moment. I have learn that Samba4 has a remarkable changes for windows clienmts. I want to use Samba 4 to authenticate my windows xp and windows 7 clients machines. In my home based pc I am using vmware to test samba 4. For that purpose I am using CentOS 6.3 with Samab4. Kindly guide me taht it is possible that I replace windows server as a PDC controller and use samba4 to onlu authenticate my clients. I does not want Group Policy only and only authentication . So what u suggest me. Is it good to use samba 4 instaled of windows server. thanks, gardenair

    Read the article

  • Intel turbo boost - in reality

    - by gisek
    I have an Intel i7-3630QM processor in my laptop. Its speed is supposed to be from 2.4 to 3.4 GHz in turbo boost mode. In reality, will it ever run all cores on full speed (3.4GHz mentioned above) at the same time? I heard somewhere that this additional 1GHz is shared between all cores in laptops. If the boost is 1GHz per core it's pretty impressive (over 40% speed up). What does it really look like? How long can a processor run in turbo mode?

    Read the article

  • mod_jk fails to detect error state because JBoss gives 404, not 500

    - by Ilya Sher
    Configuration: Apache + mod_jk, several workers on other machines (load balancing). When JBoss fails to deploy an application for example because of failed connection to the database, requests to /myapp/somepage generate 404. How do I configure JBoss to return 500 for everything under /myapp when the application failed to deploy? Additional info: Since 404 is not an "error state" code, mod_jk does not mark the worker as failed and continues to route the traffic there. Since there might be valid requests to this application also generating 404, I can not configure mod_jk to treat 404 as an "error state"

    Read the article

  • Is anyone else using OpenBSD as a router in the enterprise? What hardware are you running it on?

    - by Kamil Kisiel
    We have an OpenBSD router at each of our locations, currently running on generic "homebrew" PC hardware in a 4U server case. Due to reliability concerns and space considerations we're looking at upgrading them to some proper server-grade hardware with support etc. These boxes serve as the routers, gateways, and firewalls at each site. At this point we're quite familiar with OpenBSD and Pf, so hesitant at moving away from the system to something else such as dedicated Cisco hardware. I'm currently thinking of moving the systems to some HP DL-series 1U machines (model yet to be determined). I'm curious to hear if other people use a setup like this in their business, or have migrated to or away from one.

    Read the article

  • Why do password entries over ssh take so long?

    - by Dean
    When I'm ssh'd into my server, any time I enter my password, there's a 40 second delay before the server responds. This occurs when logging in, as well as whenever I run a command via sudo. The delay does not happen when I run su and enter my password however. Using the -v flag for ssh doesn't show anything during this time. Looking at Wireshark, all traffic between the two machines stops while this is happening. Any idea what's happening, or advice on how to investigate this? The server is running Debian squeeze (6.0.4)

    Read the article

  • Troubles uploading big pics to Picassa when pics are on NAS

    - by Bascy
    Hi all, I'm working from Vista a laptop on a WLAN, my pictures are all on a shared (samba) network drive on a FreeNAS system, which has a wired network connection to the WLAN router (DLink 655). I'v noticed that when i try to upload pics to picassa, and select files from the network share the pics won't upload successfully. If i first copy them locally (on the laptop) and then upload them to picassa, everything works fine. If i try to copy smaller pics (<500kB) it doesnt go wrong Anybody know what is going wrong here?

    Read the article

  • OSX 10.6 integration into NIS/netgroup/automount infrastructure

    - by mdpc
    I have an existing infrastructure where accounts are maintained under NIS (yp) with no local unix accounts. Also, all the standard maps including hosts, mail aliases, netgroups, etc...are maintained in this form. Extensive use of the UNIX/Linux automounter with items scattered over the network on NFS servers. There are NO ACLs on any local or shared files. All mail needs to use basically the nullclient sendmail configuration feeding into a different system. I now have a requirement to integrate an Apple OSX 10.6 system into this environment and make it run seamlessly. My initial reading and second-hand information seems to indicate that this may not be possible on the native OSX 10.6 system. I'm concerned. Any ideas as to how to accomplish this task and make everybody happy? Thanks PS: I have never used an Apple OSX system.

    Read the article

  • Different output by df and du

    - by user1263746
    I have an old RHEL 5.5 Box, When I fire df -kh I see that /var is of 49 GB with 100 % usage. But then I go inside the /var cd /var and I fire du -kh I see that only total 300 MBs are used. I am not able to get it. Is is possible that this partition is shared with different partition? I see that /opt is also of 49 GB. I think that they are same partition. I need to free /var to start mysql but there is hardly anything to be deleted. Any pointers?

    Read the article

  • Cheapest server per gigabit throughput [closed]

    - by nethgirb
    I'm looking for a set of servers for performance testing a network, and secondarily testing some applications on the servers. Their most important task is simply to pump out data: from an application like memcached or just dumped from a large file in memory into a TCP flow (i.e., disk performance doesn't matter). This should happen over one or more 1 gigabit Ethernet ports, and the machines should run Linux (ideally), or perhaps Mac OS X or some other *nix. Other than that, there are few constraints (e.g., even something ARM-based could be fine). So here's the question: What's the cheapest server per gigabit? Price and power are both considerations.

    Read the article

  • Social-Networking Startup, Hosting Plan

    - by pws5068
    I've created a social networking community which is soon ready to release, and I'm trying to decide on a type of hosting plan. I have considered options such as VPS and Reseller plans. I anticipate (or hope for at least) a significant amount of traffic/bandwidth in the not-too-distant future. If I open a reseller, will I receive the same amount of server lag during busy hours that I do with a shared account? How significant is the profit margin with the reseller option? Aside from generalized "configurability", what advantages merit purchasing a VPS? Is there anything stopping me from reselling space on a VPS account? Features I need Include: PHP, MySql, Unlimited Domains, Ruby on Rails, Remote Database Connections

    Read the article

  • Windows Remote-App Server 2012 Office 2013 User Settings not saved

    - by dave
    I have a Windows Server 2012 with RemoteApps enabled. It's running the latest Patches etc. It has Office 2013 installed and Excel and Word are shared to all users. Now I got the Problem that after each Reboot all User Settings are lost. I have a few users who pin previously opened Documents so they dont need to remember all Paths and those are all gone after Reboot. Also last opened Documents is empty and after a Server reboot it brings the office 2013 Window for First time setup where it asks if you want to connect to skydrive and all that. In the RemoteApps Collection I enabled a Userprofile-Drive 100GB drive E: for Storing User profile data. There is a Domain of course and there is no GPO Preventing the user from Storing settings etc. We also got an older Terminal Server 2003 in the same Domain where this is not happening. Any ideas why this is happening that all the Settings are lost after Reboot?

    Read the article

  • How does iperf calculate throughput and jitter?

    - by Someone
    I've read that iperf basically tries to send as much information down a connection as quickly as possible reporting on the throughput achieved. This tool is especially useful in determining the volume of data that links between two machines can supply. is it possible to gather the same results by sending regular data, as in not testing data? what I'm trying to do is this; sending data in the foreground while in the back ground gather statistics (throughput and jitter). so can anyone tell me how iperf calculates these two values ?

    Read the article

  • Implications of allowing Windows clients to use NTLMv1?

    - by Boden
    I have a web application that I'd like to authenticate to using pass-through NTLM for SSO. There is a problem, however, in that NTLMv2 apparently will not work in this scenario (without the application storing an identical password hash). I enabled NTLMv1 on one client machine (Vista) using its local group policy: Computer-Windows Settings-Security Settings-Network Security: LAN Manager authentication level. I changed it to Send LM & NTLM - use NTLMv2 session security if negotiated. This worked, and I'm able to login to the web application using NTLM. Now this application would be used by all of my client machines... so I'm wondering what the security risks are if I was push this policy out to all of them (not to the domain controller itself though)?

    Read the article

  • One-way forest trust between geographically distributed forests using Server 2008 R2

    - by bwerks
    Hi all, I'm planning out a joinder between two domains, as would take place with contracting companies. Forests A and B exist in distant sites, and there is to be a one-way forest trust so that domain users in Forest A can be authenticated on machines in Forest B. In order to facilitate this, each forest's domain controller must be able to contact each other in order to set up & confirm the trust, but my question is what underlying networking magic must take place beneath it. So far the prevailing approach has been to maintain a VPN connection between the two sites, but the technet documentation seems to indicate that DNS forwarding may be the way to go. Is this the case? Furthermore, if DNS will suffice, does that mean that there must be a server running DNS on boundary servers in each domain so that they can be reached from across the internet? How must they be configured? Thanks!

    Read the article

  • Cannot connect to xdebug over virtual network - Vagrant Virtualbox

    - by smix96
    I'm trying to set up a development box using Virtualbox / Vagrant / Puppet with the intention of eventually provisioning all my machines up to production. However this is stopping me going forward at the moment. I've installed ubuntu lucid over windows 7. When trying to connect to xdebug by starting a debugging session in eclipse, it hangs at 57% (common in eclipse when it cannot communicate with xdebug). Here is my xdebug.ini and the settings here are appearing in phpinfo(). xdebug.remote_enable=On xdebug.remote_handler=dbgp xdebug.remote_host=192.168.56.1 xdebug.remote_port=9000 I'm now wondering if it's an issue with port forwarding? If eclipse is looking on port 9000, will it find port 9000 on a virtual machine? Any help would be greatly appreciated!

    Read the article

  • Cloud Computing - Multiple Physical Computers, One Logical Computer

    - by Koobz
    I know that you can set up multiple virtual machines per physical computer. I'm wondering if it's possible to make multiple physical computers behave as one logical unit? Fundamentally the way I imagine it working is that you can throw 10 computers into a facility one day. You've got one client that requires the equivalent of two computers worth, and 100 others that eat up the remaining 8. As demands change you're just reallocating logical resources, maybe the 2 computer client now requires a third physical system. You just add it to the cloud, and don't worry about sharding the database, or migrating data over to a new server. Can it work this way? If yes, why would anyone ever do things like partition their database servers anymore? Just add more computing resources. You scale horizontally with the hardware, but your server appears to scale vertically. There's no need to modify your application's infrastructure to support multiple databases etc.

    Read the article

  • Cloud Computing - Multiple Physical Computers, One Logical Computer

    - by bundini
    I know that you can set up multiple virtual machines per physical computer. I'm wondering if it's possible to make multiple physical computers behave as one logical unit? Fundamentally the way I imagine it working is that you can throw 10 computers into a facility one day. You've got one client that requires the equivalent of two computers worth, and 100 others that eat up the remaining 8. As demands change you're just reallocating logical resources, maybe the 2 computer client now requires a third physical system. You just add it to the cloud, and don't worry about sharding the database, or migrating data over to a new server. Can it work this way? If yes, why would anyone ever do things like hand partition their database servers anymore? Just add more computing resources. You scale horizontally with the hardware, but your server appears to scale vertically. There's no need to modify your application's supporting infrastructure to support multiple databases etc.

    Read the article

  • Why is windows 7 marking all my extra RAM as "hardware reserved"?

    - by pinkmuppet
    I have a new HP Z800 Workstation, running Windows 7 Enterprise 64 bit. The machine has 4GB of RAM -- windows reported 1.99GB available, 2GB was "hardware reserved". I added another 2 GB and windows reports that there is 6GB of physical RAM, but now 4GB is "hardware reserved". The BIOS sees the full RAM amount as well. The video card is an nVida Quadro FX 580, so it shouldn't be using any shared memory at all. whats going on? Z800 Specs are here.

    Read the article

  • Upgrade AirPort on Macs to support Snow Leopard's Wake on Wireless/WLAN?

    - by wojo
    Snow Leopard now supports Wake on WLAN, but not all hardware supports this. For example, my Octo Mac Pro from early 2008 has an AirPort card, but it does not support this. Nor does my 2007 2.33GHz MacBook Pro. For reference to what is needed, look at http://www.macrumors.com/2009/08/28/a-closer-look-at-snow-leopards-wake-on-demand-feature/ which includes a screenshot of what the System Profiler should show. It's pretty hard to find Apple parts, but is it possible to put newer cards into these machines to have them support Wake on Wireless?

    Read the article

  • Best security practice for small networks - wifi, lan,

    - by Grimlockz
    We regularly setup small networks for clients in different locations to allow them to work on different products now the question what should be the best security practice. Currently we have a wifi enabled with WPA2 and most laptops connect to this but some will connect to a cabled switch connecting to the router. We are thinking on what we should do to increase the security on our small networks - We do have have security on the laptops so you can share directly to the other persons drive by a simple Windows user account. Some suggestions are: We get a LAN switch with ACL control and mac filtering for the hard wired connections? We get acl working on the wifi via a good Cisco router? ipSec policies on all machines? IP filtering and fixed IPs? I suppose people are worried that anyone can plug into the switches and get the access to the network . Summary: Maintain a level of decent security that can be replicated easily to every setup that we do for clients

    Read the article

< Previous Page | 241 242 243 244 245 246 247 248 249 250 251 252  | Next Page >