Search Results

Search found 5643 results on 226 pages for 'machines'.

Page 127/226 | < Previous Page | 123 124 125 126 127 128 129 130 131 132 133 134  | Next Page >

  • Get the "source network address" in Event ID 529 audit entries on Windows XP

    - by Make it useful Keep it simple
    In windows server 2003 when an Event 529 (logon failure) occures with a logon type of 10 (remote logon), the source network IP address is recorded in the event log. On a windows XP machine, this (and some other details) are omitted. If a bot is trying a brute force over RDP (some of my XP machines are (and need to be) exposed with a public IP address), i cannot see the originating IP address so i don't know what to block (with a script i run every few minutes). The DC does not log this detail either when the logon attempt is to the client xp machine and the DC is only asked to authenticate the credentials. Any help getting this detail in the log would be appreciated.

    Read the article

  • How to use sshd_config - PermitUserEnvironment option

    - by laks
    I have client1 and client2 both are linux machines. From client1: client1$ssh root@client2 "env" it displays list of ssh variables from client2. Things I did on client2: I want to add new variable to client2 . So I edited sshd_config to PermitUserEnvironment yes and created a file environment under ssh with following entry Hi=Hello then restart sshd /etc/init.d/sshd Now from client1 trying the same command client1$ssh root@client2 "env" didn't provide the new variable "Hi". ref: http://www.raphink.info/2008/09/forcing-environment-in-ssh.html http://www.netexpertise.eu/en/ssh/environment-variables-and-ssh.html/comment-page-1#comment-1703

    Read the article

  • FoxPro 2.6 DOS on Windows 7 64-bit

    - by Rolando
    I support a company that has a very old, mission critical, FoxPro for DOS 2.6 (FPD) application. For variuos reasons the company didn't adapt/migrate their app, which, ironically, has been running even better under Windows XP (and 32-bit Win7) because the OS allowed new features like more reliable networking, distributed printing, email integration. Unfortunately for this company, most new machines now come with a 64-bit version of Windows 7, which is incompatible with their FPD app. I know this time the writing is on the wall: the only long-term solution is to migrate their app. But I wonder if anyone can suggest a temporary alternative path, which doesn't involve either: downgrade 64-bit Windows to 32-bit, or run the app on a virtualized 32-bit XP

    Read the article

  • Should I anticipate any problems trying to use the same SSL Cert on 2 computers (primary, backup)?

    - by Matt
    We have a production machine running IIS6 with a wildcard SSL certificate. The certificate that was installed is not exportable. We want to upgrade the system to IIS7. As part of this venture, we're creating a backup/failover server that will serve the exact same websites - when we take the primary down for upgrade, the secondary will take over. As such, the secondary also needs the SSL certificate. However, since the certificate was not exportable, this means re-keying it from Go Daddy. Per http://help.godaddy.com/article/867, I know that by re-keying the certificate the original will stop working. I'm still pretty new to SSL certificates, so are there any problems I should anticipate when installing the same SSL certificate on 2 different machines?

    Read the article

  • How do you manage updates without a staging environment: CentOS 6.3

    - by Gregg Leventhal
    I am managing about 20 servers, many of them virtual. They are almost all different purpose, and none are clustered. I have a distributed LAMP stack, a few application servers, some build servers, a few KVM hosts. They are CentOS 6.3 mostly with a few Ubuntu (unfortunately). I don't have the resources to setup a staging environment where I can have duplicates of my machines and test updates before rolling them out. I am taking file backups. What I want to know is how you are approaching backing up your Linux systems. I assume you don't just do yum update, but then how are you choosing the packages worthy of updating? When (if ever) are you updating the kernel, etc.. How do you test updates without a staging environment? Snapshot and hope for the best?

    Read the article

  • A Linux DHCP server that will listen on an non-broadcast (tap) interface?

    - by TomOnTime
    Are there any Linux DHCP servers that will listen to what Cisco calls an "unnumbered" interface, or what others might call a "NBMA" (non-broadcast) interface. I have a Linux system that connects to a number of others using GRE tunnels. The machines on the other end send DHCP requests to this machine, I can see them with tcpdump. However, ISC DHCP 3.0.3 refuses to listen to the interface because it is non-broadcast. The interface I'd like DHCP to listen to is: tap2 Link encap:Ethernet HWaddr removed-for-privacy inet6 addr: removed-for-privacy/64 Scope:Link UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 RX packets:518 errors:0 dropped:0 overruns:0 frame:0 TX packets:510 errors:0 dropped:0 overruns:0 carrier:0 collisions:0 txqueuelen:500 RX bytes:196242 (191.6 KiB) TX bytes:52425 (51.1 KiB)

    Read the article

  • How to synchonise address books across multiple PC's

    - by Rob
    I am using Thunderbird 15.0.1 on two different Windows computers (Vista and Win 7). I maintain the email list for a local club, and often need to send out emails to all members from my home and office Pc's. I use a Thunderbird Address Book to maintain a list of club members and would like any changes made to that list to synchronise to both machines. Can anyone recommend a good way of doing this? I've found a couple of extensions that claim to do this, but both seem to have review that suggest they no longer function correctly since they are quite old extensions. Is there another tool that I would be better off using? Rob

    Read the article

  • 'server was unable to allocate from the system nonpaged pool' error

    - by cop1152
    Currently users at my company are unable to access their home folder (or any folder on our domain controller). I am working remotely and cant physically access the machine until tomorrow morning, but I can access it with RDC. When I attempt to ping other machines on our network from the machine in question (connected with RDC), I immediately get the error PING: transmit failed. error code 1450. Event Viewer is full of 2019 errors: The server was unable to allocate from the system nonpaged pool because the pool was empty Does anyone have experience with this specific issue? I have found some info with Google, but would like to hear what the pros at SF have top say. This is a Windows 2000 server.

    Read the article

  • Virtual environment firewall with CSF + iptables rules on VM?

    - by luison
    We are getting into virtualization with a Proxmox VE (OpenVZ + KVM) server. Our plan for firewall is to have CSF (http://configserver.com/cp/csf.html) running on the host machine as we've had a reasonable good experience with it in the past. Apart from that we plan simple firewall rules on the VM machines (mostly OpenVZ containers with same kernel) and maybe fail2ban simple specific rules. I would appreciate comments with anyone with similar experiences? I understand all traffic comes via the host machine so a combined firewall there with specific firewalling on the VM should work, alltough some iptables rules are hard to get to work on OpenVZ containers.

    Read the article

  • Access points fighting for dominance?

    - by Phillip Oldham
    We have a small office with a large number of wireless devices (a mixture of desktop machines, laptops, and wifi-enabled phones) all working from a single Apple Airport Extreme which extends our wired network. I've added another Airport Extreme for resiliency, since we've been seeing a decrease in performance and (as far as I understand) access points can only handle a small number of clients. I set the new AP to extend the current network so that the clients weren't constantly switching between different wireless networks, however as soon as this AP was configured all the wireless devices started seeing network trouble, flicking on and off. I'm assuming that this is because both APs are reasonably strong, and the client can't decide which to use. What is the best route to follow to resolve this? What I'm aiming for is wireless resiliency; preferably having two APs share the network load, or if this isn't an option then having a primary AP with a "fail-over", should the primary go down for any reason.

    Read the article

  • Correct CPU Frequency in BIOS

    - by akula
    One of the machines I have is a 10 year old one. I can't discard it due to some sentiment (Mom). Legacy board: Mercury 810e, 133 MHz FSB Processor: Pentium III Tualatin 1.2 GHz Observation: I see an entry in the BIOS for the CPU Frequency from 6.0 to 11.0. Last value is Safe Mode. I don't know what value to choose for my CPU. So I'm running that CPU in "Safe Mode". What is the correct value for this CPU? Is Safe Mode really safe to run this CPU?

    Read the article

  • linux/solaris kill many proccess with one command

    - by yael
    Is it possible to kill all find process with one command? I do not want to kill each process as kill -9 25295 , kill -9 11994 , etc.. Rather, what I want is a simple way or command that kill all find process (my target is to perfrom this action on linux and solaris machines). $ ps -ef | grep find root 25295 25290 0 08:59:59 pts/1 0:01 find /etc -type f -exec grep -l 100.106.23.152 {} ; -print root 11994 26144 0 09:04:18 pts/1 0:00 find /etc -type f -exec grep -l 100.106.23.153 {} ; -print root 25366 25356 0 08:59:59 pts/1 0:01 find /etc -type f -exec grep -l 100.106.23.154 {} ; -print root 26703 26658 0 09:00:05 pts/1 0:01 find /etc -type f -exec grep -l 100.106.23.155 {} ; -print

    Read the article

  • OpenLdap 2.4 on centos 6 doesn't listen on port 636

    - by Oliver Henriot
    I have an openldap 2.4 server on centos 6 whose confg I copied from those I have running under openldap 2.3 servers on centos 5 machines. On openldap 2.3, specifying TLSCACertificateFile, TLSCertificateFile and TLSCertificateKeyFile with correct values makes the server listen on port 636. This is not the case on the openldap 2.4 setup. I have configured it with loglevel -1 but I have not seen any clue as to what might be wrong and reading the openldap 2.4 manual doesn't indicate if any of the other TLS related parameters are now mandatory. I don't think so though because if I run the service manually, using "# /usr/sbin/slapd -u ldap -h "ldap:/// ldaps:/// ldapi:///"", the server does listen on port 636 and I can query it using "ldapsearch -H ldaps://myserver:636". Is there something I am missing to get the server to listen on port 636 without having to always launch it manually? Is this linked to centos 6 or openldap 2.4? Thank you. Cheers,

    Read the article

  • bridge to share WiFi internet connection over eth0

    - by rubo77
    In this script there is implemented that you create a mesh network and connect other machines to your device to serve an internet connection with dhcp: echo "starting bridge to share internet connection over eth0" ifconfig eth0 up promisc brctl addbr br-freifunk brctl addif br-freifunk bat0 brctl addif br-freifunk eth0 echo "internet starting, this may take some minutes due to latency..." echo "(use" echo "tail -f /var/log/syslog" echo "in another window for debugging)" echo echo dhclient br-freifunk echo "The error 'Rather than invoking init scripts through /etc/init.d...' can be ignored:" dhclient br-freifunk # without bridge: dhclient bat0 This works fine with the freifunk-mesh network. But how can I serve internet in a normal case? I would like to connect to any open Wifi I find with my network-manager or wicd (yes, with the graphical GUI) and then start a small script that creates a bridge to my network-out on my laptop I use Ubuntu and my networkcards are eth0 and wlan0

    Read the article

  • Technicolor TG582n with external DHCP server [on hold]

    - by Jack
    We have a small home setup with a Technicolor TG582n on Plusnet ISP. We have a Samba4 DC with DNS forwarding enabled. The DC forwards to the Technicolor. However, the client machines have their DNS settings manually set. This is an annoyance when using laptops on other networks. We would like to have DHCP handled on the server machine, such that when a client connects to the Technicolor, it gets its IP and DNS information from the DHCP server, eliminating the need to manually set adapter DNS settings. However, I cannot find an option to disable DHCP on the Technicolor and am not completely clear on how one would point DHCP services to the server from the Technicolor if there were the option. So, how would one make the Technicolor use an external server for DHCP leases?

    Read the article

  • How to migrate KVM based VMs running in LVM setup to Vmdk images

    - by Bond
    I am using KVM on Ubuntu Server 10.04. and Virtual Machines are running on it in LVM. I have to migrate some of them to Vmware server.How can I achieve this? I searched and came across some links but they all talked converting vmdk images to qcow or so.In this case I have OS in LVM. I also looked at man page of qemu-img and as I understand it should do what I am asking in this thread. But how exactly should I proceed in this case.Since it is not a file based image (OS running in an LVM which has filesystem in that LVM). So I am not able to understand what should I be doing to achieve the same. Can I achieve the above with snapshots of LVMs rather than shutting down the VM itself.

    Read the article

  • nfs server on cygwin slow

    - by Weltenwanderer
    The setup: We run an instance of cygwin nfsd on a Windows 2008 Server (Xeon 3,2 GHz). There are several Sun Solaris and SunOS machines accessing the shares. This is the exports file: /disk3 (rw,all_squash) /disk2 (rw,all_squash) Those paths are soft linked to the relevant cygdrive/d/path/to/dir paths. Some of the folders contain up to 10k files. The Problem: ls -la on the mounted folder on the sun boxes takes 2 - 3 minutes and the general read performance is really bad. cat filename displays the file in slow bursts and this hurts performance on tasks that access those shared files heavily. Processor load is not the issue, the nfs server idles most of the time, the cygwin tasks never get over 1% load.

    Read the article

  • What is the specific advantage of a blade server for virtualisation?

    - by ChrisZZ
    We are planning to implement a VDI Solution. We had some discussions about Blade vs Rack. As we are only planning to implement 75-100 Clients, we calculated, that we would need 2 Servers with Dual 8C Processors - and a shared storage server. This calculation is based on a paper by ORACLE, that says, 12 active virtual machines per core. Now, for buying to servers, a blade does not scale financially. But the Blade has some other advantages: a) The interconnectivity between the blades is super-fast. b) IO Virtualisation Are there other advantages, that we should consider, that would make up for price - and are this advantages so important, that we should think about investing in the blade?

    Read the article

  • Cannot schedule task to run under domain user account other than current user

    - by Filburt
    On our Win 2008 machines I can't schedule tasks for domain users because the domain name does not resolve to network name but the AD dc name. The "network name" looks like ABCDEFGE-HIJKLM and the "dc" / "name" would look like ABCDEFGE-HIJKLMN. When selecting the domain user account the account qualifier will look like ABCDEFGE-HIJKLMN\task.user. This results in an "invalid account" error. When however keeping the currently logged in user it will display ABCDEFGE-HIJKLM\current.user. Does this behaviour result from the presumable "illegal" domain name? Is there a workaround for this? update I could of course log in as the desired domain account and create the task but since this account is a account used for running services I want to avoid creating a user profile on the machine.

    Read the article

  • Arch Linux shows a blinking cursor instead of a booting installer

    - by fakedrake
    Arch Linux shows a blinking cursor instead of a booting installer. I ran sudo dd if=archlinux-2010.05-core-i686.iso of=/dev/sdb1 and checked the MD5 sums. I tried to boot it on two different PCs and got the same result: instead of booting GRUB — or anything useful, for that matter — it just showed a blinking cursor at the top left corner of the screen. The machines became unresponsive to any kind of input, the flash drive LED didn't seem to blink or shiver at all and there seemed to be no other activity whatsoever. I tried using another flash drive, but the machine completely ignored it, booting Windows "normally."

    Read the article

  • Sharing VPN from XP to Linux virtual

    - by user55513
    Hi I will keep the question simple with few details, if more is needed tell me. Thanks. I am playing around with networking and virtualbox. So I have 2 virtual machines. One is windows xp and another has linux in it. Windows XP has 2 network cards. One for the internet and the other to connect to the linux computer. Linux has one 1 card and it can only access the internet passing through Windows XP. Internet works fine until I activate OpenVPN on Windows. When I turn on VPN, Linux has no access to it. How can I make linux use VPN too? I believe it has something to do with using the "route" command on XP to send VPN traffic to Linux but I was not able to figure it out how yet.

    Read the article

  • Remote Desktop fails with no error

    - by Duane
    Recently, I've been having problems connecting to a remote machine via RDP. I can connect fine exactly once, but all subsequent attempts fail until the remote machine is rebooted. I seem to authenticate properly, but the RDP logon dialog disappears after showing, "Configuring remote session..." for a few seconds. No errors or other useful info is written either to my screen or the Event Logs of the two machines involved. I've tried disabling remote sound as others have suggested, with no success. I'm still investigating a solution, but would appreciate any thoughts. Fortunately, I can reboot the box remotely w/o causing any problems, but this is a less than ideal workaround.

    Read the article

  • Best practice for scaling a single application source to multiple nodes

    - by Andrew Waters
    I have an application which needs to scale horizontally to cover web and service nodes (at the moment they're all on one) but interact with the same set of databases and source files (both application code and custom assets). Database is no problem, it's handled already with replication in MongoDB. Also, the configuration of the servers are the same (100% linux). This question is literally about sharing a filesystem between machines so that its content is always correct, regardless of the node accessing it. My two thoughts have so far been NFS and SAN - SAN being prohibitively expensive and NFS seeing some performance issues on the second node with regards to glob()ing in PHP. Does anyone have recommended strategies or other techniques that don't involved sharding data across nodes or any potential gotchas in NFS that may cause slow disk seek times? To give you an idea of the scale, the main node initialises it's application modules in ~ 0.01 seconds. The secondary is taking ~2.2 seconds. They're VM's inside a local virtual network in ESXi and ping time between them is ~0.3ms

    Read the article

  • Router loses connection to internet randomly

    - by tvanover
    I have a Belkin wireless G plus Mimo router and it randomly loses the connection to the internet. It seems to happen for a second or two, not enough to impact day to day browsing. But enough to disconnect games, or stall videos I am watching. It happens whether I am plugged into the router or connecting via wifi. I am unsure if it is my router or Comcast that is causing this issue. If I connect a computer directly to the modem the problem never occurs, so that tells me it is probably the router. But I also had a Netgear wireless router WGR614, and it had the same problem. I don't know where to start looking for the source of the problem. What logs should I be combing through? All my machines are win 7 of varying versions.

    Read the article

  • Configuring subdomains for a machine (Win2k8) in a lan

    - by RMS
    I am currently setting up a windows 2008 server to host a website with multiple subdomains, all accessible only within the lan. also, there is no active directory. what I did is : 1 - computer name : 'web' 2 - in IIS, I added a site binding as 'site1.web' to the default web site 3 - added DNS role to the server 4 - added 'web' as principal zone in direct lookup zones (default options) 5 - added CNAME 'site1' From a client machine, in tcpip config I added the ip address of 'web' to the DNS list in addition to the ISP DNS. (client machine ip is from DHCP) Now browsing to 'http://web' or 'http://site1.web' works correctly. My question is, is it possible throught additional steps in the server to have the websites accessible without requiring the DNS config in all the client machines ? thanks in advance

    Read the article

< Previous Page | 123 124 125 126 127 128 129 130 131 132 133 134  | Next Page >