Search Results

Search found 9744 results on 390 pages for 'k means'.

Page 135/390 | < Previous Page | 131 132 133 134 135 136 137 138 139 140 141 142  | Next Page >

  • Google Apps - Can I configure a wildcard MX record and and catch all emai address for a domain

    - by Rohit
    I am using Google Apps Premier Edition. I want to create a disposable email address service and I want to catch all emails for a domain. This means that I should be able to catch all mails sent to an arbitrary userid and/or arbitrary domain and store them into a single Google Apps account. For example, in a single account I want to get all mails sent to: 1) [email protected] 2) [email protected] without requiring to do any extra configuration in Google Apps for abc or xyz. My app will download mails from this account and process accordingly. I have figured out that I could do (1) by specifying a catch all email address. Is the combination of both (1) and (2) possible?

    Read the article

  • Redundant Router and Load Balancing vs. DDoS attack

    - by colgatta
    With a small server farm at a hoster with great support and conditions, I worry about the increasing number of DDoS attacks against this hoster (not my web project, but other clients on the same location). I have booked a redundant router and load balancer as managed service with this hoster to share the load with all the dedicated servers. However, I was lost again today because another one's project was attacked with DDoS for hours :-( Each hour means hundreds of dollars loss whenever my adserver and tracking is not reachable. Even time-out advertising have to be paid by me but can not be resold to my clients without the servers being available. All the time, the servers, the load and traffic is OK and health, but no chance to keep this stable/online if the hoster is vulnerable. Anyone has ideas or suggestions how to protect - even against DDoS?

    Read the article

  • Local dedicated hosting space (own hardware)

    - by Scott
    Where can I find local dedicated hosting space for my own hardware? I know I can rent dedicated hosting from various companies online, but usually I think that means I'm renting their hardware too. I just need a space with a network connection and a power outlet. That's it. How much would this cost? What would I search for? Is it available easily? Or would it only be the sort of thing huge companies would do? I'm in the greater NYC area. It's for a project I'm working on, but the thing's loud and annoying. I'd be willing to pay a little to get it out of sight and out of mind. I don't even care too much about the quality of the network connection. I'd rather not rent other people's hardware cause it probably would cost a fortune to rent a machine like this (tons of ram).

    Read the article

  • Re-downloading accidentally deleted podcasts in iTunes

    - by Matt
    I deleted some previously deleted podcast episodes in iTunes 9 and I'd like to re-download them. (I don't want to manually re-download them because then they won't show up in the podcast's list in the iTunes Podcast section.) I've read http://blog.krisgielen.be/archives/54 as well as Way to get old podcasts in iTunes where the suggestion is to delete the iTunes entries for the podcasts I'd like to recover, then hold shift while re-opening that podcast's little triangle icon. Here's the problem though: holding shift and re-opening the triangle tells iTunes to refresh the RSS feed, and it's possible that the RSS feed no longer contains the item you want. What I'm really looking for is a means of getting the "Get" button (that normally appears on un-downloaded podcasts) to reappear on the episodes I've previously downloaded but deleted.

    Read the article

  • Video encoding for archival

    - by Jim
    I would like to archive some home videos (DV). I don't need to save them losslessly, but I would like to encode them in something high-quality. What format is both pretty indistinguishable from the original and will likely be readable 15 years from now? WMA makes me nervous, because it's only one company that makes it, and they're constantly coming out with newer formats. (VLC couldn't open my WMAs that Windows Movie Maker made.) Other things I've considered are h.264, Ogg Theora, DivX, and Xvid. I don't mind paying for something, but usually that means the format is owned by only one vendor.

    Read the article

  • Is there a way to make 7zip temporarly uncompress the whole archive when double-clicking on an exe?

    - by Gnoupi
    In WinRAR, one feature which I like is the fact that you can set it to uncompress the whole archive in a temporary place, if you double-click on an .exe file inside the archive opened in WinRAR. Typically, I often download small games, which I just want to try, without the hassle of creating a folder for it, etc. Same for archives containing an installer with its own separate files. In the 7-zip window, if I double-click an exe, it will just extract the exe in a temporary location and launch it. In the small game context (or installer), it means that it will simply fail, because it will miss required files in the same folder. So my question is: Is there a way to make 7-zip extract the whole archive in a temporary folder when launching an exe from inside the archive?

    Read the article

  • Restoring files from blueprint on command

    - by Nick
    I am setting up a server. I already have rented a machine running centOS 6 but I have run into a bit of a technical problem with configuring the server software: The server will have some files that it will try to read/write to them but what I need is a way to have a blueprint of these files and everytime the server restarts the files that it used get deleted and replaced by the blueprints. I have heared of a RAM disks or Virtual File Systems but didnt quite understand how they work or how to set them up. The server software is written in java which means bash commands can be run from it. I cannot modify what happens when the server shuts down entirely what I can do is run a command before the server runs the final shutdown save

    Read the article

  • Windows VPN - NO internet access

    - by sharru
    I host a network of servers behind a Fortigate 200a firewall in the DC. I connect to those servers via a VPN connection. The problem is that when i connect to the VPN, I lose my internet connection on the local PC (windows 7). I would like to be connected to the VPN and still surf the web. i guess this means to only forward a range of ip to the VPN connection. I've read other answers on serverfault, talking about "un-check the 'Use default gateway on remote network' option in your Windows 7 PPTP network connection settings". When i do that , i get internet access but no access to the servers in the VPN. Any idea how to get both working? Should i change something on the fortigate 200a config? Do i need two networks cards? Is there a place in windows to define ip range for the vpn connection?

    Read the article

  • Windows Server 2003 guest on Xen frequently stops responding

    - by smokris
    I'm running a Windows Server 2003 guest instance in Xen 3.x. This DomU runs fine for a day or two, then stops responding — I don't get any network response, and I can no longer connect to Xen's VNC console for this DomU. xm list shows this: Name ID Mem VCPUs State Time(s) Domain-0 0 6508 8 r----- 1161159.4 [A working Linux DomU] 1 512 1 -b---- 68711.1 [The hung Windows DomU] 5 512 1 ------ 67234.2 [Another working Linux DomU] 3 512 1 -b---- 163036.4 (What does the ------ mean? The xm manual explains what each of the six states mean, but not what no-state means.) If I xm destroy and then xm create the Windows DomU again, it boots right back up (with the Windows alert The previous system shutdown at [...] was unexpected.), and then stops responding after another day or two. I checked /var/log/xen/*.log, but no log messages are generated at the time the server stops responding. How should I proceed in troubleshooting this?

    Read the article

  • About Eucalyptus Virtualization Support

    - by rammayur
    I am having hp 540 notebook with processor core 2 duo. I have installed CLC,CC,SC and Walrus runs fine when I try to run NC, I get error "Not virtualization enabled" means my laptop does not support virtualization. I got it. Then, In my college, there is Dell desktop PC having processor Core 2 Duo. But it has virtualization enbled.So I get confused. How same thing has different effects. So my questions are Which thing is essential for Virtualization?? What should I upgrade in my Laptop Processor or buy a new one?

    Read the article

  • tcp handshake failed.client send rst (after syn-ack). can any one advice?

    - by user1495181
    architecture: 2 linux computer connected . on the second (192.168.1.1) one run apache server . I have a small program that take tcp packets from nfqueue change the dst ip to 192.168.1.1 in case that the dst ip is 192.168.1.2 (i know that i can do it with iptables , but my program will do more things in the future), fix check sum and return to the queue. if i call to telnet 192.168.1.1 , means that my program dosnt need to do any manipulation, handshake is OK. If i call to telnet 192.168.1.2 , my program change the dest. server get the syn and return syn-ack, but right after getting the syn-ack the client send rst. Can anyone advice? wireshark of the telnet tcpdump of the telenet above

    Read the article

  • importing old .eml files int outlook 2010

    - by Duncan Kenworthy
    I was having so many problems with my laptop & Windows vista that I have updated to Windows 7 64bit. Because I was having so may problems I wanted a 'clean' upgrade so didn't want to carry over any programs or settings. Prior to changing over I saved all the folders I wanted from MS mail to a separate drive and then updated to Win 7 and Office 2010. To my horror I don't seem to be able to import my .eml messages into Outlook which means all my mail is in limbo. Please can anyone help me to simply import my old inbox messages into my new inbox as I have given up trying using the import/export menu as there doesn't seem to be a choice that works. Help!! Not too good a nerd!

    Read the article

  • How to change Winamp's default skin to Bento?

    - by TomatoSandwich
    If I install Winamp through automated means (e.g. Using http://ninite.com/), it is automatically installed using the default 'boxy' skin. I can't work out how to interface with this skin as easily as the new 'Bento' skin. In addition, there doesn't seem to be an option to download the Bento skin, or change the skin from Default to Bento without uninstalling and reinstalling it again. Is there a menu option or a downloadable skin that is the default Bento skin? Or do I have to re-install to get this 'new default' skin?

    Read the article

  • Circumventing a manual HTML login page for "unclassified" websites

    - by auramo
    The IT department just made my life a little bit harder again: they introduced a manual HTML login page for all websites they have not "classified". This means that all the applications which try to access unclassified websites for e.g. downloading plugins do not work. Examples: Eclipse plugin installation, Maven builds etc. What would be the easiest workaround for this? The best I've come up with is try to extend/customize Ruby's httpproxy.rb that comes with Webrick. I would automate the manual login process whenever that login response page is detected. This sounds quite painful, and I think there might/should be simpler options?

    Read the article

  • How to setup dhcp3-server to advertise the DNS server the server itself has got from DHCP?

    - by Ivan
    The Ubuntu 10.04 server has eth0 Internet interface configured by means of an ISP's DHCP. At the same time the server has static eth0 LAN interface to which it provides masquerading (NAT) and LAN-internal DHCP service (dhcp3-server). As far as I've understood the manual, I had to hardcode DNS servers to advertise through LAN DHCP with option domain-name-servers in dhcpd.conf. But what if the ISP changes his DNS server IP silently (we use a SOHO-class ISP, so this won't surprise me much)? Can I configure dhcpd to advertise the DNS server the server uses itself, the one gotten by its DHCP client mechanism?

    Read the article

  • tcpdump dns output codes

    - by tim
    Captured on the nameserver: 21:54:35.391126 IP resolver.7538 > server.domain: 57385% [1au] A? www.domain.de. (42) What das the percent sign in 57385% mean? As far as I can see 57385 is the clients sequence number, a plus would mean RD bit set. Second question: what does the ARCOUNT do in the query? As I understand the tcpdump man page the [1au] means tcpdump treats this as a protocol anomalie - as would I. I see this in a lot of queries.

    Read the article

  • Private Git repo using Smart HTTP with LDAP authentification

    - by ALOToverflow
    I've been crawling the interwebz and getting my hands dirty for the last few days, but I can't seem to make it all work together. I managed to get a HTTP repo working with Ubuntu 10.04 over Smart HTTP (pull and push over HTTP) for a single repo. This means that I do the initial setup over SSH to the server (git init --bare) and after that the clients can pull and push to it (git clone http://servername/allgitrepos/repo.git). Unfortunately it's impossible to add a new repo without SSHing to the server and adding it manually) i.e. git push http://servername/allgitrepos/repo2.git (allgitrepos is available for everyone to read-write and execute) would fail talking about git update-server-info (which seems to be a general error message). So far the repository is anonymous, so I would like to authenticate using LDAP and also use the LDAP creds to make the git commit. So, how can I push new repos to the server and how can I use the LDAP creds to make the git commit. Thanks

    Read the article

  • Why some recovery tools are still able to find deleted files after I purge Recycle Bin, defrag the disk and zero-fill free space?

    - by Ivan
    As far as I understand, when I delete (without using Recycle Bin) a file, its record is removed from the file system table of contents (FAT/MFT/etc...) but the values of the disk sectors which were occupied by the file remain intact until these sectors are reused to write something else. When I use some sort of erased files recovery tool, it reads those sectors directly and tries to build up the original file. In this case, what I can't understand is why recovery tools are still able to find deleted files (with reduced chance of rebuilding them though) after I defragment the drive and overwrite all the free space with zeros. Can you explain this? I thought zero-overwritten deleted files can be only found by means of some special forensic lab magnetic scan hardware and those complex wiping algorithms (overwriting free space multiple times with random and non-random patterns) only make sense to prevent such a physical scan to succeed, but practically it seems that plain zero-fill is not enough to wipe all the tracks of deleted files. How can this be?

    Read the article

  • Database backup regardless of backup made through a control panel?

    - by developer
    I know that all CMS or CMC platforms have some sort of walkthrough for their users to fully backup and restore the database or the whole website. While we all can perform such backups (and there are even plugins which automate the whole procedure) and restore them in necessity (such as when we migrate to a new server), one can also backup the whole website by means of such control panels as Directadmin or Cpanel. Now, I just want to know if it is necessary for us to do database or website backups the way they are described by a specific CMC or CMS developer even after we perform a whole website backup in a control panel such as DA? An example of such CMS platforms is moodle. Moodle Docs describes how we can backup and restore moodle in here. So do we, out of necessity, have to make the backup the way described, or we can simply do it the Control panel way? Thanks

    Read the article

  • SSL 3.0 warning in Chrome on Ubuntu 10.04LTS

    - by Leopd
    I'm running Apache2 with SSL on Ubuntu 10.04LTS. Chrome gives me this annoying warning when I inspect the certificate: The connection had to be retried using SSL 3.0. This typically means that the server is using very old software and may have other security issues. The relevant part of the apache config looks like: SSLEngine on SSLCertificateFile /etc/ssl/... SSLCertificateKeyFile /etc/ssl/... SSLCACertificateFile /etc/ssl/... SSLProtocol -all +SSLv3 +TLSv1 The last line I added to try to address this problem, but it's not working. Any advice on properly enabling TLS?

    Read the article

  • How to defend agains botnet http requests

    - by Killercode
    I have a server with WHM + CPanel and 5 of my costumer got infected with zbot. This means that the domains they have are constantly receiving requests to certain destinations. I tried to use mod_security but seems that it can't filter every requests... I don't really know why? I still see in the access log the connection comming in and it's consuming a LOT of bandwidth and server load Those accounts have already been clean so all of those requests go to error 404 (the ones catched on mod_security I am dropping the connection). Is there anymore ways to defend against this requests?

    Read the article

  • Why is the size of windows off by 226x238 if defined via the Window Rules?

    - by Bobby
    I have installed Sawfish 1.8.2 from source on my new Ubuntu 12.04 installation following the Debian instructions, but I had this problem also with the stock 1.5.3. Whenever I define dimensions in the Window Rules for a window, the size is off by exactly 226x238 pixels, which means that 100x100 turns into 326x328. That's a very odd behavior, given that Sawfish is saving and loading the dimensions of the windows correctly (if saved via the window menu). Some additional system information: $ uname -a Linux Dagon 3.2.0-24-generic-pae #39-Ubuntu SMP Mon May 21 18:54:21 UTC 2012 i686 i686 i386 GNU/Linux $ sawfish --version sawfish version 1.8.2 nvidia proprietary driver, 9600GT Two monitors, 1920x1080 + 1440x900 in one session. Positionng the windows is working fine, only the dimensions are off by that odd number. Does somebody have an idea why?

    Read the article

  • How full is too full for mechanical hard drives?

    - by Sunny Molini
    I have heard many claim that it doesn't matter how full a drive is until it starts cutting into temp and virtual memory space. This doesn't make sense to me, given the nature of how the data is transacted on a hard drive. The inside of the platter presents less data per revolution than the outside of the drive does, by significant factors. The inside 40% of the radius of full size hard drive is used for the spindle, so only the outside 60% is used for data storage, but that still means that the inside track of a hard drive presents data 60% slower than the outside track. By my calculation, a Hard drive that is only 10% full should perform about 2.25 times faster than a hard drive that is 90% full, assuming that the flow is constrained by other factors. Am I wildly off base here? For all the drives I know, even the top speeds of the first 1% of the drive would be well within the bandwidth provided by a SATA 2 connection.

    Read the article

  • How to change the terminal to SCO compliant in Ubuntu

    - by kaushik
    I have a server in which SCO 5.x is installed. I have several thin clients which were supplied from our head office. Those thin clients have linux pre-installed in them. When we tried to connect to our SCO server, we have found out that the thin client does not support SCO terminal type. So the program (Which is written in Oracle forms) is not working correctly. Lots of function keys stopped working and screen also not getting refreshed correctly. Hence we decided to use Ubuntu Live USB keys in thin clients. Now the problem is how can we change the terminal type to SCO in ubuntu ? Means I want it in the text terminal (Alt+Ctrl+F1,F2...) (Not in xterm...graphical terminal emulators). Does ubuntu support SCO terminal type ?

    Read the article

  • how to have files created by CMS have the same ownership as SSH user

    - by Cam
    I am having difficulty on our ubuntu server whereby I have an SSH user that when I create files using this user the ownership is web_user:www-data The problem is when a file is uploaded or created using a content management system like joomla. When files are uploaded through Joomla - such as components / modules... The ownership is set to www-data:www-data This means that I need to then chown all new files to web_user:www-data so we can edit the files. Is there a way to set for a directory and sub-directories that all new files created have the ownership of web_user:www-data? Do I need to use something like setuid or setgid? Any help would be greatly appreciated.

    Read the article

< Previous Page | 131 132 133 134 135 136 137 138 139 140 141 142  | Next Page >