Search Results

Search found 22998 results on 920 pages for 'supervised users'.

Page 574/920 | < Previous Page | 570 571 572 573 574 575 576 577 578 579 580 581  | Next Page >

  • User account automatically filling up with dead.letter file

    - by jeroen
    I have one user account on a server with about 400 accounts that is filling up automatically. The dead.letter file in the users home directory automatically grows until the account is full (about 10 - 40 Mb per day). The user is using Microsoft Outlook to send and receive mail. What can be causing this and how can I avoid it from happening? Right now I have an emergency cron-job to delete the file but I would like "real" solution. Edit: The server version is Red Hat Enterprise Linux ES release 4 (Nahant Update 4) Edit 2: It seems mainly spam and I see different mailer headings (from php to Outlook Express) and a frequent appearing header is [email protected] Update: I have asked the hosting provider where I use that dedicated server to look into the problem as well, as it's their Control Panel that could be a cause of the problem.

    Read the article

  • Downloading file from FTP using cURL

    - by Josiah
    I'm trying to use a cURL command to download a file from an FTP server to a local drive on my computer. I've tried curl "ftp://myftpsite" --user name:password -Q "CWD /users/myfolder/" -O "myfile.raw" But it returns an error that says: curl: Remote file name has no length! curl: try 'curl --help' or 'curl --manual' for more information curl: (6) Could not resolve host: myfile.raw; No data record of requested type I've tried some other methods, but nothing seems to work. Also, I'm not quite sure how to specify which folder I want the file to be downloaded to. How would I do that?

    Read the article

  • Team City Backup from Rest API doesn't backup

    - by ChrisFletcher
    I'm using the REST API in Team City 6.5 to request a backup is created as follows: http://teamcity:8111/httpAuth/app/rest/server/backup?includeConfigs=true&includeDatabase=true&includeBuildLogs=true&fileName=filename as specified in the documentation here: http://confluence.jetbrains.net/display/TW/REST+API+Plugin#RESTAPIPlugin-DataBackup The problem is that it simply returns "idle", as does the backup status call and no backup appears. I can backup fine from the web interface and I can return a list of users from the REST API without issue also. I'm starting to suspect there is some sort of permission or config option but I can't find one.

    Read the article

  • Social Media Aggregator, Global Update via Powershell

    - by deanjmiller
    Does anyone know of a way to interface with a Social Media Aggregator using Powershell. For Instance, I would like to update my global status on digsby using Powershell. Digsby would then fan the message out to Facebook, Myspace, Twitter, Etc.. I am open to using any Social Media Aggregator that can do this.. Digsby, Seesmic, Ping.fm TweetDeek, etc.. If any of these programs have a com interface or something like it I'm sure who ever implements this first will have a large gain in users.

    Read the article

  • Should I limit end-user gigabit ports to avoid saturating uplink/trunk connections?

    - by Joel Coel
    We have a campus with 16 buildings and older 850nm 1Gbps fiber links between the buildings, that all come to a core switch for our servers that also uses 1Gbps ports. We're finally starting to replace our aging 10/100 end-user switches, and much of what we're looking at are 1 Gbps units. My question is, since the trunk/uplink lines are still 1Gbps, if I were to install 1 Gbps switches for end users, should I limit the ports to 100Mbps until I can also upgrade the trunks to avoid allowing a bad-behaving host to saturate a trunk line (since we're a college, we have plenty of mis-behaving hosts) and thereby create a DoS situation for a building, or will TCP congestion control typically take care of that for me? What if we have a lot of UDP traffic (games, video chats, even a small amount of bittorrent)?

    Read the article

  • Make backups of Dropbox folder every week

    - by ilansch
    I have a Dropbox folder which is shared by couple of users. I would like to make a backup of this folder that will occur every week and store this backup on another hard drive. I can simply copy the entire folder each time and this will be the backup, but I would like to copy only the files that have been changed or created during that week. I thought of creating a batch script that will check each file in the Dropbox folder recursively and see its modified date. If that date is later then a given one (current backup date) it will copy the file to a folder named BackUP[Date]. Do you think this solution is OK?

    Read the article

  • How to schedule time-of-day upgrades

    - by Richard
    Hello, I'm responsible for about 30 Ubuntu computers at a private K-8 school. We have only a 3Mbps internet connection serving the entire campus, and I would like to ensure that updates are done in the middle of the night - so that daytime tasks are not slowed down. I'm using Ubuntu 10.04, and have set all computers to download and install security updates via the update manager. I have also installed cron-apt, and modified the config file to stagger the start times of the upgrades from about 10pm to 4am local time. HOWEVER - this morning I arrived at the school at 7:30am and all the computers were busy downloading a large security based update. Needless to say, all internet activity was slowed to a crawl (for the next 2 hours), and the computer users were very very upset. This was the event I'm trying so hard to prevent. It seems that my scheme to ensure middle of the night downloads failed, and I'm not sure why. I've also tried some schemes using unattended-upgrades & crontab, but there always seemed to be something scheduling upgrades to occur in addition to the ones I try to force at middle of the night. Is there a sure fire way to absolutely positively guarantee that updates will occur only at one specific time? It would be nice if the update manager just had a drop down menu to specify a designated time. Thanks in advance for any help you can give me.

    Read the article

  • Tomcat access logs - are failed requests included?

    - by Maxim Eliseev
    We have a RESTful web service (Java, hosted in Tomcat on Ubuntu on Amazon EC2). From time to time it fails (not every week). When it fails, Java CPU consumption goes to 100% and it takes all available memory. It does not finish by itself. I have to restart the server. There is nothing suspicious in Tomcat access logs. I guess one of our users could submit a very "heavy" request which brought the server down. Is it possible this request is not in Tomcat logs since it never finished?

    Read the article

  • Du Meter Log file

    - by Jack
    Where can I find the Du Meter Log file? I tried searching C:\ProgramData\Hagel Technologies\DU Meter but the folder is empty. I also tried C:\Users\Username\AppData\Roaming and Local and LocalLow but none of them even have a Du Meter or Hagel Technologies folder. I even tried searching the temp folder but still nothing. I have a NetMeter.csv log file that I want to try and replace over the Du Meter log file cause I can't seem to find any other way to import data into Du Meter.

    Read the article

  • Access my local server by hostname or servername

    - by S.M.09
    I have a local host server hosting a few applications in tomcat which comes through a apache proxy The client or User trying to access these application has to access them like 10.XXX.XXX.XX:8080/appName OR 10.XXX.XXX.XX/appName But I want to replace the ip address with soem other name related to my applications. But I cannot go and enter the host name of the server in each users /etc/host Nor do I want to be setting up DNS. Is there another way to do this. I am using ProxyPass XXX YYY to redirect all applications of tomcat to port 80

    Read the article

  • Laptops trying to connect via wireless? Only able to connect only via wired connection

    - by Chris J. Lee
    It seems that all the wireless laptops in the office have disconnected and are unable to reconnect to the wifi after a terrible storm caused some power interruptions. Troubleshooting actions Verify wireless broadcasting is up With each laptop i ended up flushing their dns cache on Win 7 (ipconfig /flush ) Performed an ipconfig /all to see what the new ip is Made sure password was correct Reset the router (multiple times) restarted each laptop. Tried connecting wirelessly Am i missing some small detail ? Network information The router is a linksys/cisco WRVS4400N router. Supports a lot of fancy features i don't know how to use. We have 3 static ip's the conference room computer, the wireless printer, my computer Mostly windows machines, we have 3 mac users

    Read the article

  • Mac OS X 10.5/6, authenticate against by NIS or LDAP when both servers have your username

    - by Wang
    We have an organization-wide LDAP server and a department-only NIS server. Many users have accounts with the same name on both servers. Is there any way to get Leopard/Snow Leopard machines to query one server, and then the other, and let the user log in if his username/password combination matches at least one record? I can get either NIS authentication or LDAP authentication. I can even enable both, with LDAP set as higher priority, and authenticate using the name and password listed on the LDAP server. However, in the last case, if I set the LDAP domain as higher-priority in Directory Utility's search path and then provide the username/password pair listed in the NIS record, then my login is rejected even though the NIS server would accept it. Is there any way to make the OS check the rest of the search path after it finds the username?

    Read the article

  • Web Filter For Multiple Networks

    - by Rob
    I have been using a Barracuda web filter 310 in our network and I have just had enough of it. It does not support trunking and we have several networks that have users that need to be web filtered. (I guess if everyone just did their flippin jobs I would not have this issue) but the management wants me to get it resolved. Does anyone know the top five web filters that are better than the barracuda web filters that support network trunking so that I can have multiple domains and subnets going through it? Thanks in advance - everyone on this site is gold in my book!

    Read the article

  • where is this SFP problem occuring

    - by Saif Khan
    Hi, Inbound mail to a specific user is been forwarded to his external mailbox at another domain. Add mails are routed correctly with the exception of one domain. When mail are sent to users of my domain, including that user, the mails are delivered to everyone except him...as the mail admin I am getting the message The following recipient(s) cannot be reached: usermailbox on 1/10/2011 1:27 PM There was a SMTP communication problem with the recipient's email server. Please contact your system administrator. email.speedimpex.com #5.5.0 smtp;587 [email protected] sender domain does not match SPF records I am sure this has nothing to do with my mail server, however, can someone confirm?

    Read the article

  • meta.stackoverflow.com has a problem

    - by asker
    Sorry, this is off topic, but fact evolved that meta.stackoverflow does only allow posting with openid despite stating possibility of post per nick/email. Posted here because underlying prob stemmed from serverfault. So here is a copy: Despite stating that submission via nick/email were possible, required fields are not given. Please fix or state that critique be only issued non anonymously. Tags OpenID Login Get an OpenID Oops! Your question couldn't be submitted because: must include one of these tags -- bug feature-request discussion support users with less than 99 reputation can't create new tags. The tag 'limit' is new. Try using an existing tag instead. name and email, or your OpenID, are missing

    Read the article

  • Distributing application updates using SCCM 2007

    - by theraneman
    Hi all, If there are any System Center Config Manager (SCCM) users out there, please clarify my doubt here. I have used the ConfigMgr console to distribute a custom application to a client machine. Now I need to distribute some updated files of that application. Isnt's it possible to add those files in the same package source used earlier and advertise again? Or should I use the SCCM software update section for this? Not sure if its only me, but the Software distribution process looks much easier than the Software Updates process in SCCM 2007. Please do let me know if there any online tutorials which explain how to update a custom application. Any help much appreciated.

    Read the article

  • Online File Sharing that acts just like LAN shared drives, etc.

    - by Dayton Brown
    Hi All, Have a small business client that wants to move their current file share to the web. Specs are as follows, 20 to 30 GB of space, file sizes are normal (nothing more than 50 to 100 mb) 3 users ideal solution would be exact same functionality as windows explorer. CHEAP!!! But not super cheap. I would like to keep it around $20 per user per month. I've explored a bunch of solutions, but they are all a bit on the complicated side. Thanks in advance for the recommendations.

    Read the article

  • Problem with usb wireless mouse

    - by aiacet
    Recently I have started having problems with my wireless mouse (Wireless Optical Mouse MI-4910D). Sometimes when I start my PC or during a game the pointer/arrow stops moving. When the PC boots the pointer is locked in the center of the screen. If I'm lucky it helps to change the USB wireless adapter from port 1 to port 2 but sometimes this trick doesn't work and I have to restart my PC to get the mouse to work again. Like you can see in the product web page this mouse don't have a driver but only a tool to solve some problems. Thank you in advance to all the "super-users" that reading this question would be help me Ajax

    Read the article

  • cannot unset env variables from script

    - by w00t
    Hi, I am trying to unset all environment variables from within a script. The script runs fine but if I run env it still shows all the variables set. If I run the command from CLI, it works and the variables are unset. unset `env | awk -F= '/^\w/ {print $1}' | xargs` Have any idea how to run this from a script? Also, have any idea how to source /etc/profile from a script? This doesn't work either. I need to set variables with same names but different paths, depending on the instances my users need.

    Read the article

  • Automatically cycle numerous or large files to the trash

    - by minameismud
    I've been tasked with fixing a vendor's program that, under certain conditions, dumps gigs of junk files into a log directory. It ends up filling users' machines. My task is to figure out how to make it stop without any source code or additional running processes, and without making the program kasplode. In other words, I'm looking to use a feature of the file system to control the growth. One idea I had was to make a hard link from that folder to NUL, as you might with /dev/null in the linux world. However, my attempts to use the mklink program to create a junction result in a message that says Local volumes are required to complete the operation. Any ideas on how to complete the junction, or other ideas to solve the problem?

    Read the article

  • DNS security (hijacking?)

    - by Jongsma
    I am hosting my website on Linode and am also using their DNS/naming servers. (ns1.linode.com etc.) It occurred to me that I never have had to authenticate that the domain is mine when I added it to the domain to the DNS manager, or at any other point. I now wonder whether it would be possible for other Linode users to 'hijack' my domain by simply adding the same domain zone and pointing it to their own server. I wouldn't know how Linode could determine which are the real/authentic records. How can I be sure this doesn't happen?

    Read the article

  • Windows server 2003 mapping home drive wrong

    - by Sandman2010
    hey all, first question... we have around 30 servers in an Active Directory environment with 600 student computers and 100 staff desktops with XP SP2/3, the win server 2003 has the staff home drives on a NAS and in the last few days after some server updates is now mapping home drives to the \servername\home instead of \severname\home\%username%, its simple to re map the network drive but is annoying. we dont use login script to map home drive but use a VB script for other network drives and if we add the home drive mapping to that it works, but shouldnt the profile option in users AD account map that correctly? which do you all recommend, AD profile mapping or VB Script mapping Home drives? thanks Steven

    Read the article

  • This weird behaviour from cronjob

    - by The DOCTOR from TARDIS
    I have set the crontab like this: */5 0 * * * /www/permitChat.sh and the /www/permitChat.sh is this: # We are setting the name of file # in the variable along with complete path. sFilePath=`date +\/www\/ChatLogs\/%Y\/%m/%d_%m_%Y.txt` # First we set its permissions to # readable by all users, and then # modify them to be writable by only root. chmod a=r $sFilePath chmod u+w $sFilePath ls -lh $sFilePath The trouble I am facing is, the cron gets executed after 12:00 PM everyday, instead of executing at 12:00 AM to 01:00 AM, every 5 minutes. What could be wrong? All my system variables appear to be synced.

    Read the article

  • Deleting entire lines in a text file based on a partial string match with Windows PowerShell

    - by Charles
    So I have several large text files I need to sort through, and remove all occurrences of lines which contain a given keyword. So basically, if I have these lines: This is not a test This is a test Maybe a test Definitely not a test And I run the script with 'not', I need to entirely delete lines 1 and 4. I've been trying with: PS C:\Users\Admin> (Get-Content "D:\Logs\co2.txt") | Foreach-Object {$_ -replace "3*Program*", ""} | Set-Content "D:\Logs\co2.txt" but it only replaces the 'Program' and not the entire line.

    Read the article

  • How long do uploaded files stay in the tmp folder in Linux Ubuntu?

    - by Jean-Nicolas Boulay Desjardins
    I am building a web application where my users will be able to upload files. After the files are uploaded I need to send the files to two other servers, and after they will be deleted from the server where they were just uploaded to. I am wandering is it a good I idea to keep the uploaded files in the tmp/ folder the time the uploaded files are sent to the other two servers or should I move them to another folder incase they get deleted? I am also wandering because I would like to know if I have to build a cron script to get rid of the files that have been transfered to the other servers so that I get my disk space back.

    Read the article

< Previous Page | 570 571 572 573 574 575 576 577 578 579 580 581  | Next Page >