Search Results

Search found 12497 results on 500 pages for 'linked servers'.

Page 337/500 | < Previous Page | 333 334 335 336 337 338 339 340 341 342 343 344  | Next Page >

  • Activating ssl on tomcat

    - by toom
    I want to encrypt the http traffic on a tomcat instance via ssl. Therefore I followed the most simplistic approach described on various webpages. But anyway it simply does not work. Here is what I did: "keytool -genkey -alias tomcat -keyalg RSA" and I enterd "changeit" as the password (since this is the defaut chosen by tomcat) Altering $CATALINA_HOME/conf/servers.xml by uncommenting the following line Connector port="8443" protocol="HTTP/1.1" SSLEnabled="true" maxThreads="150" scheme="https" secure="true" clientAuth="false" sslProtocol="TLS"/ Restarting tomcat Entering https://localhost:8443 does not work. However, I can still access the page via normal http like http://localhost:8080 The logfile does not contain any suspicious information. What is going wrong here?

    Read the article

  • Can a RAID disk setup crash if only 1 hard disk fails?

    - by Steve Rodrigue
    I am a web developer. I have not much experience in hardware. For this reason, I use managed servers. This morning, one of the drives in our setup failed. However, the full site went down. I asked my web host what happened and he replied that the hard disk failed in such a way that the RAID controller couldn't work properly. Do you guys ever seen that before? Is it possible? Thanks for any help on this guys. I need to know if my web host is honest with me. Steve

    Read the article

  • Poor disk performance with high disk capacity usage

    - by GoldenNewby
    I've heard numerous times in the web hosting industry that using "too much" disk space on a drive is bad for performance. Is this just a myth? Can someone explain why this is an issue, even in a situation where the amount of IO done to the drive would be the same at 10% as it would be at 90%? I'm especially curious in the case of virtual servers. If I set up 10 Logical volumes as the virtual disks for some VMs, is it going to run better if I "waste" 20% of the disk space?

    Read the article

  • What would Stack Exchange's yearly expenses be if it were to be using a third party host?

    - by abel
    StackExchange manages it's own servers, as it should, but if SE were to be hosted on a 3rd party "cloud" hosting (like Amazon's), what would it's monthly / yearly expenses be(keeping everything else the same)? A detailed answer comparing it to the bills that Stackexchange boots currently (including power/property/staff) would help. (PS: I know that the blog is a good resource. I also understand that managing your own hosting is almost the same as setting up a hosting company and using it for your own needs. Plus is this a question for meta or does it fit within serverfault's purview?)

    Read the article

  • ruby on rails server is intermittently slow

    - by Richard
    My rails installation was chugging along nicely. Last night we had to perform a hot-patch with was really a standard deploy of some exception code. Once capistrano finished the operation one of our admins discovered that there were two long running passenger processes. While we have deployed release over the past two weeks it would appear that these processes have been here and alive the whole time. Granted they could have been zombies or any other artifact and at this point we do not know what state they were in. Which leads me to the question: There are so many moving parts between the rails application and the OS/hardware that being a SME is probably no longer possible. So; how does a sysadmin perform root-cause analysis with any certainty? And: When do I just start rebooting servers?

    Read the article

  • Hyper-V R2 as a VM on another virtual OS

    - by Tim
    I am trying to perform Microsoft Platform testing for a vendor application. The problem I have is that it requires the test be done on Windows Server 2008 as a VM on Hyper-V R2. Currently, I have access to a virtual server with just Hyper-V and also have access to an ESXi server. The crazy idea (which may just show how little I know about how virtualization technology works) is to install Hyper-V R2 as a VM onto one of these other servers. Then create a VM for Windows Server 2008 on this Hyper-V R2 VM. I can not just upgrade the current Hyper-V server as the VMs currently running would need to be taken off-line and are system critical (and I don't have rights to perform this upgrade). Has anyone tried this? Will this even work?

    Read the article

  • what does the 'm' unit in munin mean?

    - by nbv4
    I'm using munin as a tool for monitoring my servers. On some of the graphs, the units are marked with a 'm'. For instance, my apache accesses graph is labeled 100m, 200m, 300m, along the y-axis. What does the 'm' mean? I understand 'M' (caps) is mega as in megabytes, the 'k' is kilo, the 'G' is giga, but what about 'm'? At first I thought it was million, but theres no way apache is serving 100 million accesses even per decade.

    Read the article

  • Hosting solution for sensitive client data

    - by Mark
    Hello, We are developing a web application that will deal with highly sensitive (financial) data of clients (audience is medium to large sized businesses). Clients will be under scrutiny from regulators & auditors and, as such, we will be too. More importantly to give clients a level of comfort our application and related hosting arrangement should instill a lot of confidence with them. We are looking into using a cloud based service like Linode, Amazon EC2, etc. To allow for maximum flexibility We are keen on putting everything on virtual servers and avoiding having to buy our own hardware. Does a cloud based service make sense for our particular scenario? If not what type of hosting should we consider? If so what should we look out for? Thanks!

    Read the article

  • Explaining svn / apache permissions error? (I know "how" but not "why")

    - by Neil
    I have the following error occurring on occasion when trying to do an svn switch (have it set up to do via a web request): svn: Can't open file '/root/.subversion/servers': Permission denied This happens after an apache httpd.conf change and corresponding restart. How to fix this? I can get it to fix by doing an apache restart - BUT, it often takes multiple tries. Curious if anybody can explain this. Why did this error go away on my 8th apache restart, but not on the prior ones (with no edits to the conf file)? Basically, I kind of have a "how" in terms of solving this, but I don't have a "why" . . . Thanks!

    Read the article

  • Firewall blocks FTP PASV response

    - by harper
    I have an FTP server that supports passive server mode (using PASV command). This works fine with Windows XP. When I want to access this server from Windows Vista or Windows 7 with firewall enabled I experience an immediate connection shutdown. A reset packet is sent to the server, and the socket is signaled that the server has reset the connection (which is not true). The problem disappears when the firewall is disabled. Connections to other FTP servers work correctly. The difference is that the server's response to PASV does not enclose the address field with parentheses. This is legal as documented in RFC-959 and RFC-1132. How can I configure the firewall to stop this bad behavior?

    Read the article

  • Exchange 07 to 07 mailbox migration using local continuous replication

    - by tacos_tacos_tacos
    I have an existing Exchange Server ex0 and a fresh Exchange Server ex1, both 2007SP3. The servers are in different sites so users cannot access mailboxes on ex1 as from my understanding, a standalone CAS is required for this. I am thinking of doing the following: Enable local continous replication of the storage group on ex0 to a mapped drive that points to the corresponding storage group folder on ex1 At some point when the replication is done (small number of users and volume of mail), say on a late night on the weekend, disable CAS on ex0 (or otherwise redirect requests on the server-side from ex0 to ex1) AND change the public DNS name of the CAS so that it points to ex1. Will my plan work? If not, please explain what I can do to fix it.

    Read the article

  • Exchange 2010 allows outside access to network files

    - by user2891127
    One of our users discovered by accident he could access our network files from his smartphone while at home. No VPN needed. He was sent an email with an internal link to a network share on his android. When he opened the email and clicked on the link, he could browse our files while at home. Looking at the access logs, the connection to the share and files he accessed came from our mail server (Exchange 2010). We have no sharepoint servers running at all, and certainly not on the Exchange server. What is this function/feature called, and is it possible to turn this function/feature off? Should I turn this off?

    Read the article

  • Public key repository - does this exist?

    - by allegroconmolto
    This is one of those "Surely this already exists, and if not, I damn well better build it!" moments. Here's my problem: I run a devteam, of around 10 individuals. Each of us has our own private key(s). When I go set up a new server for us, I have to import all of their keys to it. And if we start working with a contractor, I have to get his or her key too, and then import it and place it in .ssh/authorized_keys on the server. Ideally I'd like to be able to do something along the lines of: import_key allegroconmolto And it would contact a public repository of public keys, see if there was a user named allegroconmolto, and if so, import their key and add it to my authorized_keys. Adding future users would then be as simple as running import_key for them. I could then scp the authorized_keys file to all my servers and be done! This seems so obvious that I feel like it surely must exist somewhere.

    Read the article

  • Is it possible to track down who or what changed a shared permission?

    - by user45574
    Today I received an email from one of my users asking why he couldn’t access his shared folder on one of our servers. Example: \\servername\share\ = access denied. When I checked the share permissions on the folder I was surprised to see that the user had been removed from the "shared permissions" list. Now my question is: Is it possible to track who or what deleted the users share permissions on the folder? I have studied the different event logs, but couldn’t find any indication of anyone who had changed the share permissions. Kind Regards Martin

    Read the article

  • Performance Drop Lingers after Load [closed]

    - by Charles
    Possible Duplicate: How do you do Load Testing and Capacity Planning for Databases I'm noticing a drop in performance after subsequent load tests. Although our cpu and ram numbers look fine, performance seems to degrade over time as sustained load is applied to the system. If we allow more time between the load tests, the performance gets back to about 1,000 ms, but if you apply load every 3 minutes or so, it starts to degrade to a point where it takes 12,000 ms. None of the application servers are showing lingering apache processes and the number of database connections cools down to about 3 (from a sustained 20). Is there anything else I should be looking out for here?

    Read the article

  • KVM guest storage difference with NBD and NFS

    - by WojonsTech
    I am setting up my own little private cloud for my own use maybe for a project or to. I am using linux kvm on debian 6. I have 3 servers 2 of them for compute nodes and 1 storage node. I would I have already installed kvm made a few test machines got my networking setup. I have 2 nics on each server 1 nic is for web traffic other nic is for network traffic. My first Idea was to use NFS for storing the guest machines which can range in size, maybe 8gb maybe 100gb, it just depends. I was doing have heard of nbd before seems like it could work but I dont know what the performance differences are and if it will effect my enviroment, nfs looks like it will be easier to use.

    Read the article

  • Is it possible to change User's Home Directorys permission in OSX?

    - by Sosiska
    Most of your staff uses OSX as main operation system. The problem is that recently we were attacked with some odd malware: users are getting zip-file via mail, and when they open this zip file, they execute a binary keylogger malware, that is inside this zipped file. (One click is enough). We have some non-technical limitations and due this limitation we can't configure user's mail servers. But actually we have physical access to their laptops. As far as I know, there is possible to mount user's home directory without "x" (execution) permission in Linux and *BSD. So users can't run some binary file inside home directory. Is it possible to configure OS X so that user can't execute files inside /Users/?

    Read the article

  • Dependency diagramming / mapping tool [closed]

    - by Lars
    I am looking for a tool that allows me to easily create and maintain dependency maps of our mission critical servers, apps, processes, etc. It needs to be intuitive and easy to work with and be able to generate diagrams that clearly show the dependencies graphically. What would be some good tools for this? I have looked at videos for AssetGen Sysmap and BluePrint from Pathwaysystems.com, and they both seem to fit my needs, but there has got to be more good systems like them that I can look at. I want to make sure I pick the best system for our needs (and limited budget).

    Read the article

  • Straight-forward to Virtualise with XEN or KVM on IBM Server System x3650 M4 791562G?

    - by ChrisZZ
    I want to built a virtualised Server Environment using XEN or KVM. The virtual machines should be purely debian systems - so XEN or KVM should be a sane choice. Now while buying servers, I am confronted with the fact, that the vendors obviously only support commercial solution. I think, on a good server, one should be able to install uncommercial software as well - but of course sometimes systems have hardware, that requires drivers, that are not found in the OS Community. So I am asked the question: Is it straight-forward to use Debian with IBM Server System x3650 M4 791562G with Debian - or even virtualising the IBM Server System x3650 M4 791562G using XEN or KVM. I am sure there will always be a way to achieve this goal - but this way might have a high milage - so I am not asking, whether this is theoretical possible, but whether this should be straight-forward and practically easy to do, no major headaches to be expected.

    Read the article

  • Can a DNS server return NXDOMAIN and an unresolved CNAME even though the alias exists?

    - by Martin
    I'm trying to understand why one given DNS server is not resolving our domain correctly. When I look up www.foo.com from this name server, a CNAME foo.bar.cc is correctly returned but there's no attached A record and the response status is NXDOMAIN. I understand that this could happen if the DNS server can not resolve the alias foo.bar.cc. However, if I use the same DNS server and look up foo.bar.cc directly, I always get the correct A record. When using any other DNS servers I've tested (different ISPs, Google DNS, Open DNS) our domain resolves correctly. Given this situation, is there some possible problem in our DNS that could cause this? Or does it seem like this given DNS server is to blame? The DNS server in question is used by a lot of people and it seems more likely that the problem is caused by us, but I don't see how - any clues?

    Read the article

  • Can't connect to memcached

    - by DMClark
    We currently have memcached running on CentOS. None of our PHP applications can connect, have tried multiple applications trying to establish access. The most informative PHP error we get is: "Memcache::get() [function.Memcache-get]: Server 127.0.0.1 (tcp 11211) failed with: Permission denied (13) in /var/www/.." memcached 1.4.5 PECL 2.25 We can telnet and it works. IP tables is full access from lo to lo. We've tried this on two different servers with both compiled version and the rpm in CentOS 5.5 and get the same result. Is there anything fairly obvious that we are missing?

    Read the article

  • How should a small team using multiple OS's deploy over github?

    - by Toby
    We have a small development team that have recently moved to using github to host our projects. The team consists of three developers, 2 on Windows and 1 on Mac. I am currently researching the best way to deploy applications to our Linux servers (dev and production). Capistrano running locally would be ideal but from what I read this won't work for Windows machines. It looks like the best way is to use a post-receive hook in github, I can see how this would work for auto deploying to dev, but I don't see how we could then deploy to live. I have found paid projects like http://www.deployhq.com/ but it feels like something that a quick bit of code should be able to do for free, I just can't seem to get myself pointed in the right direction! I was wondering what would be considered best practice for small team deployment involving multiple local OS's and github.

    Read the article

  • High IO latency on ESXi 4.1 on HP DL580 G7

    - by teddydestodes
    at my Company we were experiencing some weird spikes on IO latency on one of our ESXi instances. we've spend 24h figuring out whats wrong and no clue so far. after giving up we put all the disks into a different server (HP DL380 G7) with much less RAM and only one 6(HT) cores (from 12 on the DL 580) which run fine for about 2 hours. I dont know the specs for the DL380 but both servers have a Smart Array P410i with BBWC (the DL 580 has 1GB) Is it possible that one(or all) of the disks is failing without actually failing?

    Read the article

  • Website ocasionally does not load on first click

    - by tfe
    Today I noticed that my website hosted on a virtual server ocasionally does not load on first click. I click on some link, browser starts loading page but nothing loads and does not not appear any error message (like "connection reset by peer etc). Nothing. When I click the same link again, page loads immediately. The same situation on 2 computers in different browsers. It happends not always, maybe on each 20... or 30 click. Sites from other servers load without this problem. Any ideas what can cause this problem?

    Read the article

  • Is there a Utility that will scan selected locations and return all files older than a certain date?

    - by CT
    Can anyone recommend a utility that can scan specified directory locations (network shares specifically) and return all files older than a certain date? I am looking to implement a data retention policy at my workplace. As our amount of data grows it puts a large strain on our backup routines. I would like to move old data to some sort of archival system. Extra points for the ability to move queried old files to another location for archival and the ability to create schedules for when this occurs. Many thanks. EDIT: Windows Shop. Mostly Windows 2003 Servers.

    Read the article

< Previous Page | 333 334 335 336 337 338 339 340 341 342 343 344  | Next Page >