Search Results

Search found 17470 results on 699 pages for 'single quote'.

Page 388/699 | < Previous Page | 384 385 386 387 388 389 390 391 392 393 394 395  | Next Page >

  • Windows failover cluster - virtual node with multiple Client Access Names

    - by mclaassen
    I recently encountered a customer environment in which they had failover cluster where one of the virtual nodes had more than one 'Client Access Name' (i.e. more than one IP address and DNS name for the single virtual node). Long story short we had to modify our software to deal with this situation, but we want to recreate the situation in house to test it before releasing. I have been unable to locate any information about how or why you would end up with a virtual node that has more than one access name. Does anyone know how I can set up a Windows failover cluster where a virtual node has more than one access name/IP?

    Read the article

  • Git Repo to mantain the app configurations in several servers

    - by user62904
    Hi! I need to versioning in a GIT repository, configurations of a particular platform, spread across multiple servers. Take into account that in each of these servers there are completely different configurations, while the application is the same. What is the best way to do this? Create a branch for each server repository.git:conf -- [branch Server 1] repository.git:conf -- [branch Server 2] repository.git:conf -- [branch Server N] Note: This method seems to me, that is difficult to maintain because each change in the server configurations, I need to create subbranches which becomes confusing. Create a single repo with a different directory for each server repository.git:conf/Server 1 repository.git:conf/Server 2 repository.git:conf/Server N Note: This is easy to mantain Create a repo for each server repository_1.git:conf repository_2.git:conf repository_N.git:conf Note: This method requires me to create a branch for each new server. There are other methods, what are the best practices in this case? Should I use the one that I feel most comfortable? Tks, Gulden PT

    Read the article

  • Seek first to understand, then to be understood

    - by BuckWoody
    One of the most important (and most difficult) lessons for a technical professional to learn is to not jump to the solution. Perhaps you’ve done this, or had it happen to you. As the person you’re “listening” to is speaking, your mind is performing a B-Tree lookup on possible solutions, and when the final node of the B-Tree in your mind is reached, you blurt out the “only” solution there is to the problem, whether they are done or not. There are two issues here – both of them fatal if you don’t factor them in. First, your B-Tree may not be complete, or correct. That of course leads to an incorrect response, which blows your credibility. People will not trust you if this happens often. The second danger is that the person may modify their entire problem with a single word or phrase. I once had a client explain a detailed problem to me – and I just KNEW the answer. Then they said at the end “well, that’s what it used to do, anyway. Now it doesn’t do that anymore.” Which of course negated my entire solution – happily I had kept my mouth shut until they finished. So practice listening, rather than waiting for your turn to speak. Let the person finish, let them get the concept out, give them your full attention. They’ll appreciate the courtesy, you’ll look more intelligent, and you both may find the right answer to the problem. Share this post: email it! | bookmark it! | digg it! | reddit! | kick it! | live it!

    Read the article

  • How to avoid being fooled

    - by Dacav
    I'm a big fan of the OpenSource development model, and I think that sharing information, knowledge and ideas is the best way of working in software development. Still I think that being hired for proprietary software development must not be demonized. Of course, as there's a no reward in terms of sharing there must be a bigger reward in terms of money (i.e. I surrender all my rights for cash). It may happen that one gets hired piecework, for a single project: in this case one is more vulnerable to dishonest employers. This didn't happen to me personally, but some friends of mine had bad experiences, and lost a lot of time without being fairly retributed. Of course a contract should protect both parts. But contracts can be very generic in the specification. Software is not anything but a palpable good! Besides I don't think that contracts can distinguish between a well written software and a poorly written one. Note also that, in this (nasty) spirit, it's also likely that the employer cannot trust the employee! So also the employer should be protected by a dishonest employee). My question is the following: Which is, in your opinion, a good way of avoiding this kind of situation from the technical perspective?

    Read the article

  • Squid gives always tcp_miss reverse proxy

    - by JaakL
    I added installed latest squid3 in front of apache as reverse proxy. The problem is that it gives always tcp_miss, in fact I have not yet found a single TCP_HIT message in the log file, and most of the content is static. Relevant config values for cache_dir and refresh_pattern are default ones, directory /var/spool/squid3 exists and has some files/folders. I have 100+G free storage, but reconfigure gives warning "WARNING cache_mem is larger than total disk cache space!", which does not make any sense to me. I have googled a lot and seen with similar problems, but none of them has helped.

    Read the article

  • How to maximize parallel download from S3

    - by StCee
    I got a lot of images to load from Amazon S3 on a single page, and sometimes it takes quite some time to load all the images. I heard that splitting the images to load from different sub-domains would help parallel downloads, however what is the actual implementation on that? While it is easy to split for sub-domains like static,image, etc; Should I make like 10 sub-domains (image1, image2...) to load say 100 images? Or is there some clever ways to do? (By the way I am considering using memcache to cache the S3images; I am not sure if it is possible. I would be grateful for any further comments. Thanks a lot!

    Read the article

  • Windows 7 Built-In VPN - How to access network shares, printers, etc.?

    - by mbrion
    After I have successfully connected to a Windows 7 box via the built-in VPN: how do I access shares, shared printers, network appliances, etc.? So, call me bad at googling, but: I can find dozens of articles on "How to Set up a VPN Connection" and "How to connect to a VPN" for Windows 7... but I can't find a single article on how to access resources after connecting. I have a home VPN set up in Windows 7; I was able to connect to it from my friend's Windows 7 machine last night. I expected to be able to UNC into my shares; I also expected to be able to go to "Start Computer" and Click "Network" on the left side, and then see all of the devices on my home network (while the VPN was connected). Am I missing something obvious? How do I make this happen?

    Read the article

  • How can I use '{}' to redirect the output of a command run through find's -exec option?

    - by pkaeding
    I am trying to automate an svnadmin dump command for a backup script, and I want to do something like this: find /var/svn/* \( ! -name dir -prune \) -type d -exec svnadmin dump {} > {}.svn \; This seems to work, in that it looks through each svn repository in /var/svn, and runs svnadmin dump on it. However, the second {} in the exec command doesn't get substituted for the name of the directory being processed. It basically just results a single file named {}.svn. I suspect that this is because the shell interprets > to end the find command, and it tries redirecting stdout from that command to the file named {}.svn. Any ideas?

    Read the article

  • Multiple public/private key pairs for the same user

    - by bruceb
    First, sorry if this question has already been asked/answered - I've searched but perhaps I haven't recognised the answer.... What we have is a cluster of servers which need to access a single remote server using sftp. We are migrating from one remote server to another at the same (remote) location. We also want to refresh the public/private key pairs on the configuration as part of an ongoing security review. My question is - can we have multiple public/private key pairs for the same user between server A and server B? I want to do this to allow for cutover testing - but am concerned that the software checking keys may only try one of each type (rsa/dsa?) before rejecting the connection method and moving to the next type of key. Hope it's a straightforward question - please let me know if I need to supply more details. Thanks in advance Bruce

    Read the article

  • Does connecting to the default host via public IP from within its subnet cause any issues?

    - by username
    I'm setting up a small office network with a single public IP (let's say it's 69.16.230.117). I've configured NAT on the router with incoming traffic forwarded to the server (say the server has a private IP of 192.168.0.2). Is it okay to configure the client machines on the same subnet to access the server via the router's public IP (69.16.230.117)? In practice it's never caused me problems, but I've heard, here and there, that it is a bad idea, and one should use the private IP (192.168.0.2). Does connecting to the default host via public IP from within its subnet cause any issues? Please refrain from writing "never! it breaks the intranet!" ;-)

    Read the article

  • 1K incoming http post requests per second, each with a 10-50K file

    - by Blankman
    I'm trying to figure out what kind of server setup I will need to support: 1K http post requests per second each post will contain a xml file between 5-50K (average of 25 kilobytes) Even if I get a 100 Mb/s connection with my dedicated box (they usually give 10 Mb/s but you can upgrade), from my calculations that is about 12K kb/s which means about 480 25kb files per second. So this means I need around 3 servers then, each with 100 Mb/s connection. Would a single server running HAProxy be able to redirect the requests to other servers or does this mean I need to get something else that can handle more than 100 Mb/s to proxy things out to the other servers? If my math is off I'd appreciate any corrections you may have.

    Read the article

  • Configure a app in a subpath in nginx

    - by Rubem Azenha
    I have a nginx + passenger + single rails app setup in a Fedora server. My nginx.conf regarding my app looks something like this: server { listen 80; server_name myapp.mycompany.com; root /opt/apps/myapp_large_name/public; passenger_enabled on; client_max_body_size 4M; } So when I access myapp.mycompany.com.br it access to my rails app under /opt/apps/myapp_large_name/public. I want to be able to configure a subpath to access to another app (in this case, monit). I want something like that: myapp.mycompany.com.br/monit access the monit app. Is it possible?

    Read the article

  • unable to format external drive to HFS+ for Mac

    - by dtlussier
    I have an external hard drive (1 TB Western Digital) that is currently formatted as FAT32 and I want to reformat it to a single partition of HFS+ for my Mac. I realize that I can read FAT32 from my Mac but want HFS as it has other feathers like permissions that I'd like to have. I have tried using Disk Utility to format the drive as I have done in the past, but when I go through the process it fails and throws out an error stating that it is unable to reformat the drive to HFS. What might be the reasons that this could happen? Are there any diagnostics I could run on the physical disk to check if it is running well?

    Read the article

  • Random BSOD ntoskrnl.exe (Windows 7)

    - by nordbjerg
    I get BSOD at random times and have been for a while now on my Windows 7 machine. It is really new, and I already tried wiping the graphic card drivers and installing them again (making sure that they are of course up to date). I get a variety of bug check strings on a few different drivers. The single file that is in all of my BSODs is ntoskrnl.exe Bug Check Strings - SYSTEM_SERVICE_EXCEPTION - KMODE_EXCEPTION_NOT_HANDLED - DRIVER_IRQL_NOT_LESS_OR_EQUAL - SYSTEM_THREAD_EXCEPTION_NOT_HANDLED - PAGE_FAULT_IN_NONPAGED_AREA I would rather not resort to getting a completely new PC, as I have already thrown a lot of money on my current one. Here is a .zip file with my dumps.

    Read the article

  • Windows 7 Built-In VPN - How to access network shares, printers, etc.?

    - by mbrion
    After I have successfully connected to a Windows 7 box via the built-in VPN: how do I access shares, shared printers, network appliances, etc.? So, call me bad at googling, but: I can find dozens of articles on "How to Set up a VPN Connection" and "How to connect to a VPN" for Windows 7... but I can't find a single article on how to access resources after connecting. I have a home VPN set up in Windows 7; I was able to connect to it from my friend's Windows 7 machine last night. I expected to be able to UNC into my shares; I also expected to be able to go to "Start Computer" and Click "Network" on the left side, and then see all of the devices on my home network (while the VPN was connected). Am I missing something obvious? How do I make this happen?

    Read the article

  • Windows 7 Built-In VPN - How to access network shares, printers, etc.?

    - by Mike Brion
    After I have successfully connected to a Windows 7 box via the built-in VPN: how do I access shares, shared printers, network appliances, etc.? So, call me bad at googling, but: I can find dozens of articles on "How to Set up a VPN Connection" and "How to connect to a VPN" for Windows 7... but I can't find a single article on how to access resources after connecting. I have a home VPN set up in Windows 7; I was able to connect to it from my friend's Windows 7 machine last night. I expected to be able to UNC into my shares; I also expected to be able to go to "Start Computer" and Click "Network" on the left side, and then see all of the devices on my home network (while the VPN was connected). Am I missing something obvious? How do I make this happen?

    Read the article

  • What alternative is there to Nginx that supports http keep-alive between backends ?

    - by felace
    Hi. I recently asked a question about how to keep a backend connection persistent using Nginx, but found out it wasn't possible anyway, It is an HTTP/1.0 proxy without the ability for keep-alive requests yet. (As a result, backend connections are created and destroyed on every request.) It works all fine right now (since the connection between client and Nginx is kept alive and the result is simply the same), but I don't want to establish a new connection every single time a new request is received ,even if it's on a unix domain socket. So, what software (preferably open-source and not too tedious to configure) do you recommend to accomplish that such connections ?

    Read the article

  • Does /boot safe on top of a lvm LV (logical volume)?

    - by fantoman
    Title already asked the question. More specifically, I read in some documents that logical volumes are nice in general but not for /boot in a linux system. They say that bootloaders don't understand LVM volumes, so create a separate partition for /boot out of lvm. I recently installed Ubuntu server (9.10) for my home server, but by default /boot is created in the LVM. Everything is fine now, but I am not sure it is safe to use /boot in LVM. Second question is do I really need a physical partition (volume)(pv) for /boot or is it equally fine if I put it into a logical volume (lv) on top of a single shared volume group. Thanks in advance.

    Read the article

  • Do control groups improve system performances?

    - by qdii
    According to this website, enabling cgroups in the kernel can boost performances by sharing resources in a better way. In particular, the conclusion states that:  Nevertheless, with a little trial and error, cgroups can help you improve the efficiency of your systems’ resource usage and avoid downtime due to overusage of a single service. Kernel seeds, however, recommend to deactivate them altogether. They say: Consider these [kernel] settings poison. They remain nothing but system slow-downs. They are all off by default [in the proposed kernel config file]. Who should I trust?

    Read the article

  • &ldquo;Why do transactional messages all have the same priority?&rdquo;

    - by John Breakwell
    I answered this question on the MSMQ forum on MSDN and thought it worth sharing here. The poster wanted to know why all transactional messages have a fixed priority of zero (instead of 0 through 7). They wanted the guaranteed delivery of messages to a queue but wanted to assign different levels of priority. Some aspects of MSMQ were defined way back in the last century and this is one of them. If I remember right, the reason was to avoid the following scenario: You have a single transaction of 3 messages (a, b and c) with priorities 5, 3 and 4 respectively. The messages are sent in order a,b,c The messages arrive in the queue and are arranged in order a,c,b to reflect priority order This breaks the guaranteed order part of the transaction.  I know that very few people send more than one message in a transaction but that is a scenario that MSMQ has to be able to handle; for the majority, including yourself, this scenario is irrelevant which is why you are surprised by the absence of transactional priorities. The options, therefore, available to the Microsoft developers were to: Implement code that allowed you to send messages with variable priority as long as any messages within the same transaction were the same priority, or Define a set priority for all transactional messages As you can understand, option 1 would be a complicated arrangement with all the necessary enforcement, error handling, user education and documentation, etc. Sure, it would have been possible to implement option 1 but I expect the product group decided to invest the development time in some other aspect of MSMQ. Now, with five versions out there, it would be confusing to change how the product operates, in addition to potentially breaking exisiting systems that have been working fine for years. So, balancing cost and risk against customer demand, I would not expect this feature to ever change.

    Read the article

  • Citrix has issues resolving network shares

    - by George
    We are having this weird issue with our Citrix (version 4.5) server (sitting on Windows 2003 r2), where a couple users have issues resolving single shared network drive. We use a logon script to map all shared drives. The weird part is that of 3 shared drives, users can access 2, but the 3rd one goes to the old server (even though the logon script points to the new server). And that issues is limited to a few users. I had them log off and re-loggin to no success. It happens just in Citrix. The file server, that is being accessed, is Windows 2008 R2. Like I said we use a logon script to map the network drive. I understand I might be a little confusing, I will gladly reword the post.

    Read the article

  • Multi-machine backup solutions

    - by Paul
    I have a growing number of PCs and laptops in my home that need a backup solution. My question is 2 fold: Most commercial backup vendors appear to license by the machine, are there any products that give a license that can be used on all your machines, without a per machine cost? An alternative is to have a centrally controlled back up strategy controlled from a single PC with an attached hard drive that can back up other machines on the network. What backup software is suitable for this approach? Will this software be windows/Linux interoperable? I've searched for duplicate questions but don't see anything that addresses the multi machine/cost issue. OSs in network will be Windows (XP, Vista, 7) and Linux variants. Not pattern to when machines will be switched on.

    Read the article

  • PCI IDE Controller

    - by mercutio
    I have a suspicion that the onboard IDE controllers may not be working. Every disk I use to setup this machine reports as damaged (using a win xp installation to test, since it gets to partition setup fastest) So, I popped an IDE PCI Controller card in, to test with, but no drives are showing up in the bios now. I went into setup and changed the BIOS settings to disable onboard IDE1 and 2 to test, but still didn't detect the drives. How do I get that working? It's a standard PC with a 160GB disk and DVD Drive on master and slave channels on a single cable, if that helps. Let me know what else I need to state.

    Read the article

  • Combine WLAN with LAN

    - by fiatux
    I have D-link DSL 2750U wireless router. My PC and notebook are connected to router by Wi-Fi. They can share files or ping each other. I have a network storage unit plugged into router by an ethernet cable. Any of the computers cannot reach that storage unit or ping it. I guess my router isolates LAN network from WLAN network. Is there a way to combine them and get the router treat all devices connected to router in a single network? Thanks.

    Read the article

  • How to attach multiple ipv6 ips to eth1 on debian

    - by Noodles
    I've just got a new server with native ipv6. I want to attach multiple ipv6 ips to eth1, but the only way I can see to do so is to attach them individually: i.e. address 2607:f0d0:xxxx:xxxx::2 address 2607:f0d0:xxxx:xxxx::3 address 2607:f0d0:xxxx:xxxx::4 Is it possible to bind whole subnets of ipv6 to a single network interface on debian? My server host tells me I have 18,446,744,073,709,551,616 ipv6 addresses for that server, surely it gets to be a nightmare to manage if they all have to bound individually (plus ifconfig would look messy). Does anyone have a solution?

    Read the article

< Previous Page | 384 385 386 387 388 389 390 391 392 393 394 395  | Next Page >