Search Results

Search found 12775 results on 511 pages for 'remote validation'.

Page 374/511 | < Previous Page | 370 371 372 373 374 375 376 377 378 379 380 381  | Next Page >

  • How do you find a new software to download?

    - by user63411
    I see that the most time on computer I loose just for searching a suitable software. How you do it? what is your way to find a software? you first go on Forum?, google? Which torrent site? which P2P prog.? website, server? my problem - my way is: For example: "I need a software that have function remote and also leaves you to copy and paste (drag and drop - file manage). And after few hours searching on Google I download LOGMEIN pro2, but than I see that is just trial. so if I need to find alternative I will spend another whole day Where to go? I am not big amateur but I need better system. I need better intorudction how to find what is the suitable software for you and where you can download it?

    Read the article

  • How to deploy Windows-8 Enterprise Apps to other users?

    - by TToni
    Windows-8 (Metro) Apps can be installed using "sideloading", bypassing the Windows store in enterprise environments. In principle this is easy: Once you enabled sideloading (which is automatically done when a Win8-machine joins a domain), you can install a signed appx-Package through PowerShell with the "Add-AppxPackage" command. But there is a catch: The App is only installed for the user who executes the command and there is no "-Credentials" parameter! I can probably solve that problem in my specific scenario, where I deploy a self-developed app through TFS build to a virtual machine with a fixed demo user (by using remote powershell in combination with "Add-Job", which does take a credential parameter and because I know the given username and the password). But that is not true in an enterprise environment, where I want to distribute my App to thousands of users. Cracking all their passwords seems a bit over the top, so what would be the "correct" way to do this? I can't find any useful information from Microsoft about this, but maybe one of you already ran into this problem and solved it?

    Read the article

  • website lookup extremely slow in ubuntu

    - by ubuntulover
    Hi I have a wireless broadband connection through a router and wireless modem. Everything works fine in Windows. However, in ubuntu on the same machine, websites seem to take longer to start loading. I think the dns lookup is slow. I think https sites may be slower, as Ijust can't log in to gmail. I am also using a mercurial repo with remote origin, and it takes forever (like 5 minutes) to push one small change. I think it is because it has to communicate through https multiple times. Should I change my dns server? I've seen that I don't have these problems at my work network (they have another dns server). This happens with the IPv4 settings being automatic (dhcp). When I change it to automatic (dhcp) addresses only, and add google's 8.8.8.8 in the dns servers, it still takes forever. Why is this happening?

    Read the article

  • rsync filtering

    - by biomed
    I use an rsync command to sync two directories remote local the command is (used in python script) os.system('rsync --verbose --progress --stats --recursive\ --copy-links --times --include="*/" --include="*good_name*.good_ext*"\ --exclude-from "/myhome/mydir/src/rsync.exclude"\ %s %s'%(remotepath,localpath)) I want to exclude certain directories that has the same files that I also want to include. I want to include recursively any_dir_name/any_file_name.good but I want to exclude any and all files that are in bad_dir_name/ I used exclude-from and here is my exclude from file * /*.bad_dir_name/ Unfortunately it doesn't work. I suspect it may have something to do with --include="*/" but if I remove it the command doesn't sync any files at all. Thanks for the help.

    Read the article

  • How can you import a root certificate to a machine level store in Windows 7

    - by ReluctantAdmin01
    I have a service (Running as local system) that uses an SSL connection. Currently this connection fails because the remote host used a private CA to sign it's certificate. For previous operating systems, I used to use the certificate manager to import the CA cert into the local machine's Trusted Root certificates store. Though I can do the steps with a windows 7 machine, it seems after a reboot that the imported certificates are gone. Here are the steps I'm doing in Windows 7: Open mmc Add Certificates Snap-in for Local Machine Navigate to Third-Party Root Certification Authorities/Certificates Import Root CA Cert. The certificate seems to work fine, using internet explorer or the service to test the SSL connection works, but after a reboot it seems like the change is reverted.

    Read the article

  • Windows share through proxy

    - by jessebs
    I have a Windows share and on a different subnet. The users on the user subnet can not see the server subnet, although there is a proxy server in place on the user subnet that can. All web traffic is sent through this proxy server (squid). Is there any way to proxy a file share so the users can access it? I have tried an NTFS symlink, but I can't access the symlink through the share. The server is Windows Server 2008 R2 -- I had to stick with Windows because it is running a Remote Desktop Gateway as well.

    Read the article

  • website lookup extremely slow in ubuntu

    - by ubuntulover
    Hi I have a wireless broadband connection through a router and wireless modem. Everything works fine in Windows. However, in ubuntu on the same machine, websites seem to take longer to start loading. I think the dns lookup is slow. I think https sites may be slower, as Ijust can't log in to gmail. I am also using a mercurial repo with remote origin, and it takes forever (like 5 minutes) to push one small change. I think it is because it has to communicate through https multiple times. Should I change my dns server? I've seen that I don't have these problems at my work network (they have another dns server). This happens with the IPv4 settings being automatic (dhcp). When I change it to automatic (dhcp) addresses only, and add google's 8.8.8.8 in the dns servers, it still takes forever. Why is this happening?

    Read the article

  • Nginx works on my linux machine but is not accessible from other computers in my local network

    - by crooveck
    In my LAN network I have a server with Scientific Linux (RedHat or Fedora based distro), I've done yum install nginx but the welcome page is not accessible from other computers in my network. When I do telnet open localhost 80 and then GET / HTTP/1.0 I get some html code from nginx, so it's running for sure. But when I want to connect remotly, doing telnet open 192.168.3.130 80 I get: Trying 192.168.3.130... telnet: Unable to connect to remote host: No route to host So I assume that there is something wrong with my network settings, maybe iptables or something else? Next step, I turned off iptables: service iptables stop and it helped, now I can connect remotely using telnet. So I think, I need to fix my iptables rules. I did some googling and found this rule -A INPUT -m state --state NEW -m tcp -p tcp --dport 80 -j ACCEPT but it still didn't allow me to connect remotely when iptables is up. Can someone please help me setting a proper iptables configuration?

    Read the article

  • Forwarding sent mail through postfix

    - by D_f-r
    So I've successfully forwarded mail from my postfix server to remote Gmail accounts using /etc/aliases and newaliases. But now of course those Gmail accounts cannot send email originating from the original, aliased addresses without further configuration. I examined the Gmail configuration and found that it is indeed possible to forward sent emails via SMTP through the postfix server, but an account with a password is necessary. Even though those accounts are simply aliases, is it possible to give them passwords to facilitate replying to emails they forward? Or is an entirely different configuration set-up necessary?

    Read the article

  • Where are incoming messages queued/stored when sendmail is communicating with a milter?

    - by Mike B
    CentOS 5.x | SendMail I'd like to better understand how milter hand-offs work. We a remote MTA initiates an SMTP session, my understanding is that Sendmail hands the message data off to the milter during that session. Where/how is this message data stored? Is this all done in memory? Or is there a "queue" area where message data for milter is stored for pickup by the milter? If it helps, let's say I'm calling a milter like this: INPUT_MAIL_FILTER(`foo', `S=unix:/var/lib/foo/foo.sock, F=T, T=C:5m;S:3m;R:5m;E:5m')dnl

    Read the article

  • cygwin ssh shortcut on windows desktop

    - by Alex Berkoff
    I have multiple servers that I need to remote into. I prefer Cygwin over Putty to do so. Anyhows - the process of opening my cool Mintty window and then typing the following commands takes too long. PS - I am using a "key" authentication to these servers. First, I double Click Cygwin Terminal shortcut from my windows desktop. Then once the terminal session has booted up, from the command prompt I type the following - $ eval `ssh-agent` $ ssh-add $ ssh <username>@<servername> Please keep in mind that my 'servername' is variable. In fact I have about 10 different server names that could potentially be inserted there - Hence my need for 10 different shortcuts. I would prefer to double click on something from my desktop that will fire up my Mintty and automatically execute the above bash shell commands. Does anyone have or can recommend a nice/elegant solution to do this?

    Read the article

  • sites now not responding on port 80 [closed]

    - by JohnMerlino
    Possible Duplicate: unable to connect site to different port I was trying to resolve an issue with getting a site running on a different port: unable to connect site to different port But somehow it took out all my other sites. Now even the ones that were responding on port 80 are no longe responding, even though I did not touch the virtual hosts for them. I get this message now: Oops! Google Chrome could not connect to mysite.com However, ping responds: ping mysite.com PING mysite.com (64.135.12.134): 56 data bytes 64 bytes from 64.135.12.134: icmp_seq=0 ttl=49 time=20.839 ms 64 bytes from 64.135.12.134: icmp_seq=1 ttl=49 time=20.489 ms The result of telnet: $ telnet guarddoggps.com 80 Trying 64.135.12.134... telnet: connect to address 64.135.12.134: Connection refused telnet: Unable to connect to remote host

    Read the article

  • Server 2003 saying unlicensed

    - by nat
    I came in this morning to our windows 2003 server - running only SQL server 2005 saying it was unlicensed. When logged straight onto the box as soon as I got past the login, it popped up the click yes to license, did that and it would just loop back around to the login. It wouldn't turn off, couldn't remote to it, after a power off and reboot, it magically decided it was in fact licensed. It didn't appear to be low on disc space or other resources. Logs not showing anything out of the ordinary. Has anyone else experienced this, or might have an idea as to what just happened? Any help much appreciated.

    Read the article

  • Mac OS X multi-user thin client server (terminal server)?

    - by username
    Is there any solution out there to turn a Mac into a true multi-user thin client server? I'd like to set up a few cheap PCs with access to a couple accounts using something like VNC, but it isn't economical to buy a new server for each user or a new license for virtualized OS X Server for each user. I'm fully aware that OS X Server lets you set up users with "network home folders," and I know there's also VNC built into Mac OS X. Neither of these fit the bill (the former requires a thick client, and the latter is single-user only) UPDATE: yay, Lion! http://www.9to5mac.com/54102/10-7-lion-allows-multi-user-remote-computing

    Read the article

  • How to record audio in Windows Server 2008 Web SP2?

    - by Aurelio De Rosa
    I face into this problem and before ask here I've searched and read a lot of question but no-one seems to fit to my problem. I have a server running Windows Server 2008 Web with the SP2 and a headset attached to it using audio jack. The server has audio driver since I can listen audio. This OS does not came with the built-in audio record tool (the one I have in Windows 7). I can access physically and remotely (using RDC) to this server with administration permission. My local pc has Windows 7 if it can help. All I want to do is to know if it's possible and how to record audio form the microphone (of the server) and store that file in the server from remote connection. I tried to install several software on the server but everyone fails with a different issues (missing DLL which if I put in the system the error doesn't go away).

    Read the article

  • Can I use my old Sony Handycam TRV815 as a web cam?

    - by Optimal Solutions
    Has anyone ever used their old Sony Handycam (or maybe any older hand-held camcorder) as a web cam? As in streaming the live video from the cam to a web server (local or remote/Internet). My TRV815 has a pretty darn good picture quality for its time.. Via USB or Ethernet. It has S-Video but I think that is input. It has RCA out (white/red for audio and yellow for video) and I have used that to display to my TV. I have web cams already but was wondering if I could put this one to use also just for kicks..

    Read the article

  • How to rsync just the current folder?

    - by Michael
    I want to sync some files from a remote server to my local computer. How can I make rsync to just copy the files with a certain file extension in the directory but no subdirectories? I assumed this to be an easy task, but embarassingly I'm not getting it for nearly 2 hours. So could someone give me an example? I did various experiments with something like the following command: rsync -a --include=what? --exclude=what? -e ssh [email protected]:/test /test

    Read the article

  • Seemingly random disconnections on Windows Server 2008 R2

    - by bobblez
    So I have a box running a Windows Server 2008 R2 in my LAN, which is running, among other things, a voip server used by other people. I myself am in the same subnet with the server, while the other users are not. Now, maybe on average twice a day, all the other users get disconnected from the server (and if I take a remote desktop connection to it, I'm not able to open any web pages either). These disconnections usually last only about 5 seconds, but sometimes they may last up to a minute. What could possibly cause this? I have all the updates installed from windows update and the motherboard of the server is asus p5b deluxe (if that is of any interest). edit: Forgot to mention that my own internet connectivity stays intact while the server's does not, and both computers are behind the same router.

    Read the article

  • Git SSH RSA keys

    - by Michael
    I thought I set up my key pairs correctly -- I can do git pulls. I can do git commits. But when I do a git push, it counts objects, decompresses, then says: fatal: the remote end hung up unexpectedly. What's the issue here? I'm a super user, so it's not folder writable / readable access problems -- it must be the way I set up the encryption key pair... how do I debug this ... since git pull works?

    Read the article

  • How would I Enable FTP on Home PC (Win7)?

    - by jp2code
    In my home, all PCs connect through a small router. Some wired and some wireless. Our Media PC (HTPC) is controlled via a Media Center TV style remote, so managing files on it is tedious. I can access all of the files on the HTPC from my desktop PC, but moving a 6 GB file from \\HTPC\Folder1 to \\HTPC\Folder2 involves copying the data to the desktop and then transferring it back. If I were on the HTPC, this would likely be handled by a simple address change for the file (i.e. it would be done almost instantly). I'm thinking if I could get an FTP program to connect to the HTPC, I could simplify things ...but how do I enable that ability on the HPTC? Then, how would I go about connecting to the HTPC? Would I simply enter \\HTPC as the FTP address?

    Read the article

  • Wait for linux machine to be rebooted

    - by Theo
    I have a small script to install on my remote machine an update. I would like to reboot the machine remotely and if it is rebooted, continue with some more commands. What I currently do is: ssh root@myMachine << COMMANDS_ISSUED ###... Tasks init 6 COMMANDS_ISSUED sleep 180s ssh root@myMachine << POST_REBOOT_COMMANDS ###.... More stuff POST_REBOOT_COMMANDS Is there a more elegant way to do it? Like pinging the machine all 5 seconds up to a maximum of 4 minutes? I play with a few linux machines which have different boot up times and if my script would continue immediately after reboot, this could safe quite some time for me. (Note: I don't want to parallelize execution over all machines as I want to see for each machine if everything worked fine)

    Read the article

  • How to set up server to be accessed from mobile devices [closed]

    - by Kgrover
    I need to set up a server that will eventually be accessed from an Android application. I don't have experience with how to go about this, but I've seen MySQL servers (Amazon EC2). I would need to store data on the server remotely (again, from a mobile device) and fetch it to display. What would be the best kind of server to use? I'm guessing I would only need around 50 GB of space. Is it possible to use a network drive and set that up as a remote server with an IP address? I would need to upload and extract data through java on Android. This is my first question on ServerFault, and I'm not sure if it's the appropriate forum. If not, please redirect me.

    Read the article

  • How do you limit the bandwidth for a file copy?

    - by wizard
    I've got an old windows 2000 box in a remote location with a T1 connection and a vpn to my location. I normally use smb mounts to transfer files but now it's time to decommission the server and copy it's backups to my location. I have about 40 gigabytes (compressed) to copy. I'm prepared for it to take a long time, but I have a few caveats. I need to limit the bandwidth so terminal service connections to the site are not affected I want to be able to resume a partial transfer There are a few small files and several large files (10-20 gigabytes). I'm familiar with rsync on *nix platforms but have had bad luck with windows and I don't know that it will really keep partially transfered files. What do you use?

    Read the article

  • How to compare differences between directories (linux)

    - by Phil
    I have two directories - one from earlier backup and second from newest backup. How do i compare what changes were made to files in directory from newest backup on Linux? Also how do i display changes in for example text and php files - i'm thinking about something like revision history on wikipedia where you see old version on one side of the screen and newest version on other and changes are highlighted. How do i achieve something like that? edit: How do i also compare remote dir with local?

    Read the article

  • Prioritize file sharing capabilities in Windows Server 2008

    - by cmbrnt
    I've got a server running Windows Server 2008, and use it mainly for sharing files throughout the domain from a number of disks. It's running on VMware ESXi 4.0, in case that matters. My problem is that when I log in to the server to check user permissions etc, the access speed the files on the remote disks almost grinds to a halt. I havn't been able to measure the speeds, but I would guess it slows down to about 100kB/s as soon as I log in. This is on a gigabit network and the problems are equal for all users, even the ones connected to the same switch as the server. I've assigned 2 GB RAM to the server, and reserved it 1,5Ghz processor power. I don't have to do anything special on the server for this halt to occur. How can I make sure file sharing is prioritized on the server, so no matter what applications I'm using it will always make sure file sharing works properly? Could this be a VMware issue?

    Read the article

< Previous Page | 370 371 372 373 374 375 376 377 378 379 380 381  | Next Page >