Search Results

Search found 87891 results on 3516 pages for 'server migration'.

Page 634/3516 | < Previous Page | 630 631 632 633 634 635 636 637 638 639 640 641  | Next Page >

  • X11 tunnelling through SSH from Fedora 3 server to Windows Vista client fails to work.

    - by MiffTheFox
    Okay, I've tried using Xming and PuTTY, as well as Cygwin/X and Cygwin ssh, and it is not working. I've been able to run X applications locally, so it's not a client-side X problem. I've connected to the server using ssh -X user@server and ended up with this: local-user@client: ~$ ssh -X user@server user@server's password: Last login: Sun Jul 19 15:26:46 2009 from 192.168.100.147 [user@server ~]$ xclock & [1] 27770 [user@server ~]$ Error: Can't open display: localhost:10.0 [1]+ Exit 1 xclock [user@server ~]$ [user@server ~]$ Here's the relevant snippet of my sshd_config #AllowTcpForwarding yes #GatewayPorts no #X11Forwarding no X11Forwarding yes X11DisplayOffset 10 X11UseLocalhost yes I can't seem to find ssh_config on the client-side. /etc/ssh doesn't exist.

    Read the article

  • SQL 2005 Install completly broken on laptop.

    - by ProfKaos
    My problems started when I set up an exploratory domain controller at home. During that process, I changed the name of my laptop, from 'erisia' to 'erisia-laptop', and this resulted in me losing my admin account on SQL Server 2005 on my laptop. SQL Server seems to store the machine name as part of the user name, so now my local admin account is 'erisia-laptop\brady', which SQL server didn't recognise. I thought I'd cut my losses and just re-install SQL 2005, but the first install was aborted, so I tried a complete uninstall, followed by the new install. The latter didn't work while the former worked way too well. Now every attempt at installing SQL Server 2005 SP2 on my laptop running XP SP3 ends in sadness, with MSXML6 being the cause, because it won't install on this OS version. WTF? Does anyone know what I can do to get SQL 2005 installed? It's the only full admin DB tool I have. SQL 2008 Express that installed with VS 2010 Beta 2 doesn't allow much server admin.

    Read the article

  • can't delete some files or change their attributes on my FTP server

    - by Revolter
    I've installed a CMS on a shared host running Apache, now when I was deleting the root directory with FTP, some folders left with a "Permission denied" error and I can't change their attributes. the best explanation I've got is that the CMS installer has placed the files and has assigned its ownership to the Apache server instead of my user name. (i don't know it can be done) Ijust haven't use the uninstaller because I've lost my admin password - -" so how to delete those folders ?

    Read the article

  • Eucalyptus/ubuntu-server-11.04 - instance is running but can't access it

    - by itgorilla
    I've installed Ubuntu Enterprise Cloud on a server and I'm able to bring up an instance from a image and the instance shows that is running. I see the IPs allocated to that instance but for some reason I can't access it via SSH. euca-describe-groups shows: GROUP admin default default group PERMISSION admin default ALLOWS tcp 22 22 FROM CIDR 0.0.0.0/0 I'm on the same network as the instance so I'm sure is not an networking problem (like routers, switches etc.). Any ides?

    Read the article

  • better JSP server setup on windows?

    - by phill
    I'm currently running my jsp pages on a windows 2003 server, apache tomcat 5.5 setup. My process involved is as follows: make change to .java file compile .java file restart apache tomcat service refresh browser check c:\program files\apache software foundation\tomcat 5.5\logs This is beyond cumbersome. Is there a better development platform to develop jsp pages which makes life alot easier? thanks in advance

    Read the article

  • Install several lighttpd services in same server

    - by Pedro
    Hi, I'm running one videos streamming site using lighttpd, but at the moment the badwith of the server is at 50%, memory is ok, disks ok... But lighttpd is giving me timeouts. I think that if I had 2 or 3 lighttpd services running in the same machine I can solve this issue. How can I setup this? Regards, Pedro

    Read the article

  • Providing DNS redirection to honeypot server for known bad domains

    - by syn-
    Currently running BIND on RHEL 5.4 and am looking for a more efficient manner of providing DNS redirection to a honeypot server for a large (30,000+) list of forbidden domains. Our current solution for this requirement is to include a file containing a zone master declaration for each blocked domain in named.conf. Subsequently, each of these zone declarations point to the same zone file, which resolves all hosts in that domain to our honeypot servers. ...basically this allows us to capture any "phone home" attempts by malware that may infiltrate the internal systems. The problem with this configuration is the large amount of time taken to load all 30,000+ domains as well as management of the domain list configuration file itself... if any errors creep into this file, the BIND server will fail to start, thereby making automation of the process a little frightening. So I'm looking for something more efficient and potentially less error prone. named.conf entry: include "blackholes.conf"; blackholes.conf entry example: zone "bad-domain.com" IN { type master; file "/var/named/blackhole.zone"; allow-query { any; }; notify no; }; blackhole.zone entries: $INCLUDE std.soa @ NS ns1.ourdomain.com. @ NS ns2.ourdomain.com. @ NS ns3.ourdomain.com.                        IN            A                192.168.0.99 *                      IN            A                192.168.0.99

    Read the article

  • Is real-time or synchronous replication possible over WAN link?

    - by johnnyb10
    The company I work for is looking to implement truly real-time file replication with file locking over a WAN link that spans over 2000 miles. We currently have a 16-drive SAN setup in our east coast office. We also have an office out in Colorado that will have the same exact SAN setup. The idea is to have those two SANs contain the same exact data at all times, which will allow us to work with the same data pool, and which will also provide use with an offsite backup solution, should a failure occur on either end. We're running Server 2008. The objective is to enable users in the east coast office to work on files and have those changes be instantly updated on the Colorado SAN as well. We also need there to be file locking so that there will be no conflicts or overwritten changes if users attempt to work on the same file. Is this scenario even possible, at speeds that would make the files usable? And if so, what software would we need to pull this off? As I understand it, DFS-R does not provide file locking, so if we used that, we would need to go with a third-party product like Peerlock. But I don't even know if DFS-R is an option. Can it replicate quickly enough over a WAN link? Can any product? It seems that if we were to use synchronous replication, the programs would be unacceptably slow, as every write would have to wait for confirmation from the other end of the link. But if we used asynchronous replication, what kind of latency would we be looking at? There is a product from GlobalScape called WAFS that claims to provide "File coherence with real-time file locking, file release, and synchronization" and says that "As files are modified, changes are mirrored instantly using intelligent byte-level differencing to minimize the impact on network bandwidth". So this sounds like synchronous replication, but that doesn't even seem possible, given physical limitations such as the speed of light. If anyone has any experience with this kind of setup, or knows whether it's even possible, I'd appreciate your input and suggestions, including recommendations for software that we should check out.

    Read the article

  • SVN shared password db on Windows server

    - by Greg McGuffey
    I'd like to have a shared password-db file for several repositories on my home svn server (run under Windows). I've figured out that I need to set all the repositories to have the same realm, but I can't figure out how to just put in an absolute path to the shared password-db. I.e. the full path is something like: c:\svn.users\passwrd What do I set the password-db setting to in svnserve.conf so it can find this file?

    Read the article

  • Web server hits from Akamai IP addresses

    - by Joel
    I am getting a lot of hits to my web server from from multiple IP addresses that are owned by Akamai. This stood out to me as the requests are for resources that are no longer available. These resources should not be being requested - that pages that refer to them are marked as being non-cachable. Any clues?

    Read the article

  • What email server should I choose?

    - by DCC
    I need a secure email server installed in debian-lenny with users in a mysql table. Also users are from multiple domains. Quota should be in mysql or a global variable for all users. What are my options ? THanks in advance for your help.

    Read the article

  • Run Rails server on client's local machine for single tenancy [on hold]

    - by rigyt
    We are building a Rails application for a client that wants to run the app locally to be close to their data center for best performance. I had assumed we would use a standard web host. What requirements should we demand of the client's infrastructure..all I can think of so far is remote access for developers and a Linux machine? We don't have server admin expertise so would need to contract this, I doubt the client has the expertise. Does this sound like a recipe for disastor?!

    Read the article

  • PRoblem with EMAIL SERVER

    - by Julio Zebadua
    Hi. I follow this tutorial step by step and in the end http://wiki.mediatemple.net/w/DV:Install_Postfix_on_Ubuntu#Configuring_TLS_and_SASL_authentication and i can make mi server work correctly the imap and the pop3 never work on roundcube and squirrelmail , also i prube with a telnet. Im new i don't have idea if the tutorial is right and yes i have a VE on mediatemple and i love it but i been all day and i can mak it work this is the error i keep get it -ERR Temporary problem, please try again later

    Read the article

  • Load balancing with one server and IIS7(.5)

    - by Lieven Cardoen
    Is it possible to configure loadbalancing on one server with IIS7? What I would like is to have three applications in IIS7 (sites). One site should forward the requests to the other two sites (loadbalanced). Problem is that at a customer of ours loadbalancing is used (with virtual servers). We on the other hand do not have (yet) a virtual environment and only one buildserver. (maybe using Application Request Routing module?)

    Read the article

  • How does one make sure or even guarantee server time are sync correctly between dozens of servers across multiple datacenter on different location?

    - by forestclown
    Currently our web applications contain a logic to check if the data sent to the web server is expired or not by comparing the timestamp of the data with the date/time of the server. Everything goes will, until some dude from data center accidentally modify one of the web server date/time and causes some disruptions in our web services. My managers are of course not happy with this, and said we shouldn't use timestamp to check expiry in the first place...anyway.... Network Time Protocol is implemented, because of data centers are spread across different continents so we have one NTP server in each data center. The servers within the data center will have cron jobs to check against the time with their NTP server from the same data center. If time is out of sync it will auto update the server date/time. But then with our managers not happy with it, and think it could still easily causes the same problem. e.g. what if someone accidentally modify the NTP date/time? what if all the NTP servers are out of sync with each other? which NTP servers we can really trust? and blah blah.. So my questions are: What are the current practice to sync date/time between servers across multiple data centers or locations? How does one manages time stamp between web apps? e.g. Server A send data (contain timestamp of Server A) to Server B (compare timestamp between Server B and the timestamp from the data to see if it has expired or not. This is to avoid HTTP replay) Should we really not use timestamp check? Thanks & Best Regards

    Read the article

< Previous Page | 630 631 632 633 634 635 636 637 638 639 640 641  | Next Page >