Search Results

Search found 9362 results on 375 pages for 'direct admin'.

Page 209/375 | < Previous Page | 205 206 207 208 209 210 211 212 213 214 215 216  | Next Page >

  • NTFS Folder has no owner, no permissions, can't be deleted

    - by winnt93
    Hi, I have a folder a 2003 member server which can't be deleted. Nothing has any permissions (domain admin and running up a cmd prompt as "nt authority\system" using psexec) - always "access denied". When I do a dir /q, the owner shows as "...". I've tried takeown.exe on the folder and also it's parent. The bad folder always reports "access denied". Also tried to reset using icacls, same thing. Explorer permissons has no sharing & security options or tabs. It works fine for other folders, even in the same directory. Bit stumped to say the least. g.

    Read the article

  • Odd Language In a BIOS Message

    - by Josh
    So I started up my laptop today and was greeted with the following message (not a direct quote): The type of the AC adapter cannot be determined. This may interfere with your computer's performance. Try unplugging the AC adapter and then plugging it back in, thanks. The problem was that I hadn't fully secured the plug into the back of the computer. However, I was a little taken aback when a message from BIOS said, "thanks." Is this normal? Any chance the message was illegitimate (virus)?

    Read the article

  • Custom host file for Firefox

    - by acidzombie24
    Instead of changing my host file I'd like Firefox to think my domain is on my own server while my other browser uses the real IP address. I used to edit my host files but that forces both browsers to change the IP address. I found change host, but it doesn't appear to use the alternative host file. I also saw a comment asking when it will work on Firefox 6+. I tried Host Admin and it fails. It works but the alternative IP address must be in your host file already (which I don't want) and it lets you deselect a domain so the host file is ignored which is not what I want.

    Read the article

  • how to properly edit hosts, hostname and resolf.conf? [migrated]

    - by Firewall
    i,v been searching the internet for a real noop tutorial on the subject but could not found any direct info. on how to edit these files the proper way. i,v got a debian internet server that i use to host some personal domains and runs squid and rTorrent. the server is up and running with no problems but i am confused about a few things. lets say that i named my server (foo), my domain is (example.com) and my public IP is 95.211.133.200 now: should /etc/hostname contains: tango.example.com or tango <----- just the server name should /etc/hosts contains: 127.0.0.1 localhost.localdomain localhost 95.211.133.200 foo.example.com foo should /etc/resolf.conf contains (along with the nameservers) both: domain example.com search example.com or just the first one. are there any other files that i should edit in order to make things right? last thing, the command: domainname returns: (none) i believe it should return (example.com). what should i do to correct that?

    Read the article

  • Runas command for user without password

    - by user156558
    I want to use the runas command in CMD and from Batch files to gain elevated permissions, but I don't have an admin password on my PC because it's my home computer and I'm the only user. Every time I try to use the runas command it asks me for a password, then tells me that blank passwords are not permitted, is there any way round this? I know I can just right click and select run as administrator, but I would much rather have a command that can do it as that is not always an option, and it's just a bit of a pain. I could just be making some kind of basic mistake as I'm not hugely experienced with command prompt, but I'm not completely clueless.

    Read the article

  • Domain account credentials fails to start scheduled tasks

    - by neoco
    I have 2 Windows 2000 servers in a domain 'DC', which run numerous windows scheduled tasks under the 'DC\Task-User' account. These tasks are and have been running successfully in the past couple of years without any account/password modifications. Yesterday, the tasks went to 'Could not start' status. Myself having an admin account, was able to run these tasks with my credentials. Today, I ran the tasks again under the DC\Task-User account and they seem to be running fine without any issues. This has happened in the past too. I'm not into networking so don't know much about the underlying problem. Has this something to do with a Domain controller account authentication? Any pointers will be great!

    Read the article

  • Running 'dd' command at startup?

    - by Usman Ajmal
    Hi, I have set a script to run at Linux startup. The script contains a following line of code dd if=/dev/sda2 of=/dev/sda5 ?> result.txt Now, when my Linux Desktop appear, result.txt contain dd: opening '/dev/sda2': Permission denied If I prefix the dd command with sudo as: sudo dd if=/dev/sda2 of=/dev/sda5 ?> result.txt the result.txt contains sudo: no tty present and no askpass program specified Is there a way I can get around this problem? What I want is to copy 2nd parititon to 5th when a user logs in no matter if he is root, admin, Desktop or an unprivileged user. Thanks a lot as always.

    Read the article

  • How can I unobstrusively backup a few client's email?

    - by tladuke
    This is a small office. Our web/email server is a shared host. In the office we have a windows 2008 box up all the time that runs our NAS and a couple other services. I don't have access to the ISP admin stuff, but I assume it has cpanel or something like that. I can get access if I ask. I want to get email backed up from the server to our NAS without the users having to do anything. I suppose I could set up Outlook on that server with everyone's account, but that's a terrible idea, maybe (would sent mail. The boss uses outlook, but we have Apple Mail and Thunderbird clients too. I guess the important thing is that outlook look at the backups, so boss is happy. Then again, maybe it should be stored in whatever is the most portable format (that will work on NTFS) This is for about 10 users.

    Read the article

  • Problems restoring old backups in NetBackup 6.5

    - by gharper
    I had a server that was decommissioned & replaced last year, and since the server was no longer in use, I deleted it's client & backup policy from the NetBackup Admin Console shortly afterwards. I recently got a request to restore a file from the old server, however when I specify the source client for the restore, I get an error message saying: WARNING: server (backupserver) does not contain any backups for client (oldserver) using the specified policy type (Standard) as requested by client (backupserver). [Ok] In addition to that error, I can't seem to run a Client Backup report on the old client any more to determine what tapes I need to recall in order to re-index and restore the files... My questions: Does deleting the client somehow remove NetBackups ability to ever restore files from the old system, even if the backups have a retention period of infinity? Is there a way to restore the file from the tape, assuming I can figure out which tape I need?

    Read the article

  • Debian/Ubuntu: Enabling "dist-upgrade" behavior for unattended-upgrades?

    - by Mark Renouf
    We've got a customized distribution of Ubuntu, a repository with some custom packages and we run unattended-upgrades on a number of systems. What we want to be able to do is supply an update of one of our packages which might have a new dependency which is not yet installed. I understand apt normally prevents that from happening automatically, and using dist-upgrade would permit it. How can I get that behavior so our unattended upgrades work the same way? Ideally we'd only want new packages installed if one of our packages causes it to be needed (either as a direct dependency or a child, etc.) Should I be aware of any potential problems or increased risk of breakage. The systems are generally not easily accessed via the console so anything causing a problem requiring manual intervention would be very bad!

    Read the article

  • free -m output, should I be concerend about this servers low memory?

    - by Michael
    This is the output of free -m on a production database (MySQL with machine. 83MB looks pretty bad, but I assume the buffer/cache will be used instead of Swap? [admin@db1 www]$ free -m total used free shared buffers cached Mem: 16053 15970 83 0 122 5343 -/+ buffers/cache: 10504 5549 Swap: 2047 0 2047 top ouptut sorted by memory: top - 10:51:35 up 140 days, 7:58, 1 user, load average: 2.01, 1.47, 1.23 Tasks: 129 total, 1 running, 128 sleeping, 0 stopped, 0 zombie Cpu(s): 6.5%us, 1.2%sy, 0.0%ni, 60.2%id, 31.5%wa, 0.2%hi, 0.5%si, 0.0%st Mem: 16439060k total, 16353940k used, 85120k free, 122056k buffers Swap: 2096472k total, 104k used, 2096368k free, 5461160k cached PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND 20757 mysql 15 0 10.2g 9.7g 5440 S 29.0 61.6 28588:24 mysqld 16610 root 15 0 184m 18m 4340 S 0.0 0.1 0:32.89 sysshepd 9394 root 15 0 154m 8336 4244 S 0.0 0.1 0:12.20 snmpd 17481 ntp 15 0 23416 5044 3916 S 0.0 0.0 0:02.32 ntpd 2000 root 5 -10 12652 4464 3184 S 0.0 0.0 0:00.00 iscsid 8768 root 15 0 90164 3376 2644 S 0.0 0.0 0:00.01 sshd

    Read the article

  • NGINX - Two different rails apps under same domain

    - by Murkin
    I have two different Rails (passenger) apps that I wan to host on one server: somehost.com/ <-- App #1 somehost.com/admin <--- App #2 Tried playing with the 'location' directive, but failed to have both operate. Can someone suggest the correct approach ? (I would prefer both to share same environment, only launch from different directories) EDIT: Sample (desired) config Trying to do something like: server { listen 80; server_name myhost.com; rails_env production; passenger_enabled on; location / { root /opt/main_site/public/; } location /dev { root /opt/admin_site/public/; } }

    Read the article

  • Win Server 2008 R2 SP1 Repair Disk

    - by Horst Walter
    Is there nothing like "creating a repair disk" for Win Server 2008 R2 SP1. Or am I just missing it? Under Win7 I do just enter "Repair" and it finds "create a repair disk". Sorry - but I am not a professional Admin. Just using the server for my development environment. If there is no repair disk, is there something similar or something I should create in case I do have a system failure? Sure, a full backup is done, but anything else?

    Read the article

  • Eliminate single point of failure for webservers?

    - by George Bailey
    I know in DNS, that each of the DNS servers will be tried to see if they will respond I know in email that in the event of a failure it will go to the next one in the list or it will hold the mail for a period of time As far as I know,, in webservers,, the browser will get one of the webserver IP addresses and try it and if it fails it will give up. Is this correct? If so,, then the only way to direct traffic away from a failed IP address would be with the DNS servers.. and even that would not update immediately?

    Read the article

  • SQL Server 2008 data migration to another SQL 2008 server

    - by Confy
    I am currently running SQL server 2008 (SQLserverold) which homes databases for systemn center Service Manager 2012 Datawarehouse. This server is very old and needs to be decommisioned. Prior to decomissioning the serve, the databases have to be migrated to another clustered SQL server with two nodes(SQLServer1 and SQLserver2). I am not a DB admin no experience in SQL, and I have been given the task of migrating the Databases on the old server to the new custered environmnet. Can some point me to the right direction as to how i could do this. Step by step instruction would be helpful. Thank you in advance confy

    Read the article

  • Advice for an EC2 Architecture and Deployment Strategy

    - by Mark
    My company is currently migrating several websites and PHP web applications (standard LAMP stack) from three in-house servers to Amazon EC2. Because we had only three servers, we clustered several low-traffic websites with perhaps one high-traffic web application, and served them from the same server. The server admin has pretty much copied the previous architecture wholesale onto the EC2 instances, simply upping the instance size to account for the highest traffic client that occupies that particular instance. This architecture might be okay if it wasn't for deployment. Any time one of these sites/apps changes, it means redeploying the entire instance, along with the 30 sites/apps it hosts, instead of just updating one. How can we architect our cloud in a more modular fashion? Should each app get its own appropriately-sized instance? What is the best strategy for deployment in this type of situation?

    Read the article

  • Can't copy files from network drive

    - by user630320
    I have weird problem with copying files. When I copy file from network drive into C drive nothing happens but when I copy file from network drive to desktop I can copy the file. Also if I copy files from desktop into C it works fine. I have full local admin permission on this PC and the network drive. I have try these things Created new profile Run Windows Update Run checkdisk I'm using Windows XP 32bit pro Update: Network path: \\server1\shared\folder PC: C:\ (this doesn't work) C:\Documents and settings\Userid\Desktop (This works fine)

    Read the article

  • Can't read directory owned by my group

    - by Jonathan
    I moved the postgres data directory to a separate partition and it works great. The directory is owned by postgres user and postgres group. d-wx------ 11 postgres postgres 4.0K 2010-06-11 08:28 data/ I added myself to the group > sudo addgroup me postgres > groups me me : me adm dialout cdrom plugdev lpadmin admin sambashare postgres And gave the group read and execute permissions to everything in the directory. sudo chmod -R g+rx ./data d-wxr-x--- 11 postgres postgres 4.0K 2010-06-11 08:28 data/ But I still can not CD or LS the directory. > ls data ls: cannot open directory data: Permission denied What beginner mistake am I making?

    Read the article

  • Android : How to open youtube links on my facebook messenger app, in the youtube app? [migrated]

    - by dev_android
    I am unable to play YouTube videos from the links I get on my Facebook Messenger app. Here's what happens: A friend of mine posts a video link to me (YouTube link) as a message on Facebook. I receive it on my Android mobile in the Facebook messenger app. I click on it and it shows a Dialogue-Box (you can call that a pop-up) which asks me to choose an app to open the video with. Now earlier, there used to be the video player apps in my mobile along with the internet browsers installed on my mobile that used to come up there. So I would select an internet browser, and it in turn would re-direct me to the "YouTube" app itself. However, now, I am only getting video-players in that Dialogue Box as an option to open the YouTube video, but of course, if I select them, they are not able to stream the video. Has anyone found a way around this problem? It is really bothersome.

    Read the article

  • sharepoint 3.0 site restore/import trouble...

    - by Trondh
    Hi, We have some old sharepoint data (from a WSS 3.0 SP1 or SP2 install), that I need to restore. Problem is: This is a time management site, and one of the fields automatically pics up the user name of the user that enters data, and this is used to keep track of who worked when. Now, when I import this into my temporary sharepoint 3.0 server, these fields are blanked, and the creator of the element is replaced by my admin user (the account that ran the import job). So, to the question: Is there any way at all to grab hold of these data before the sharepoint import job "destroys" them? I'm using stsadm -o import. I don't care if I have to pick the database itself apart manually, I just need to know if it's possible to get hold of these fields with data intact from my export files... (Backup you say? It was deleted loong ago. This sharepoint export is all we have...)

    Read the article

  • how can I get rid of duplicate administrators on 2 macbook pro's

    - by BCITY
    I got a new MacBook Pro (insurance payout) and passed my old one (previous generation MacBookPro 2008) on to a staff member whose MacBook Pro died. I used carbon copy to move my data to the new laptop but couldn't do the same with her move as the older Macbook wouldn't allow it as it was on an old OS. Ended up just migrating her software onto my old MBP. Now however, they both have the same admin identities and her computer is synching with my mobile me account. Can I change the administrator identity and save all her applications? I am hoping that I don't have to wipe and reinstall everything to change her computers identity.

    Read the article

  • Password History Storage and Variability Comparison

    - by z3ke
    I believe this situation would be similar to many others out there, so maybe some of you can shed some light... Supposedly, when making password changes through MS exchange every 90 days, you cannot use any simple variation of one of your old passwords, up to whatever limit the admin's set for a system. My question: If your previous passwords are only stored as hashes, how can they check for the "just changed one letter" case. Wouldn't they have to have access to the old plain-text passwords in order to make those comparisons? The only other thing I can think of is if upon original creation of a password, they also stored all other one character permutations of it, so that they can be banned later?

    Read the article

  • SFTP over double server hop

    - by josh.trow
    I'm trying to work out a method to allow me to access files on an SFTP server than I cannot access from my local machine. Currently, I have to SSH to a remote server (it is in a certain IP block that the final SFTP server will accept from), then from there SFTP to the destination server. From there, I get the files I am interested in, thereby dropping them onto the middleman server, from which I can get the files either over a Samba share or with a direct scp. I also work in the reverse, where I drop the files onto the middleman, SSH to it then SFTP to the destination and put them into the appropriate folders. My goal is to shorten this. The unfortunate restrictions are that my machine is Windows (I use KiTTy and/or Cygwin) and I cannot modify the middleman server (or destination server) in any way. I am willing to use command line or GUI programs so long as it works and is free. Any ideas?

    Read the article

  • mod hosts file, connect to joomla on remote server

    - by Kate
    I've just acquired an account on a remote server with Joomla installed. I was instructed to add xxx.xx.xxx.xx name.ca www.name.ca to the hosts file which I found in /private/etc/ . I had to su to my admin account and use sudo to mod file and found that hosts is also found in /etc/ though it is apparently the same file. I attempted to flush the the DNS cache using dscacheutil -flushcache and then launched Safari and entered address xxx.xx.xxx.xx/administrator but got a 404 error. Joomla was set up for me by the server owner and accessed from his Windows laptop to demonstrate so I know it should work but no go here. Can anyone suggest what the problem might be?

    Read the article

  • Ubuntu apt-get install (--download-only) executed from another machine on behalf of mine

    - by Maroloccio
    I have a server on a network segment with no direct or indirect access to the Internet. I want to perform an: apt-get install <package_name> Is there a way to somehow delegate the process of downloading the required files to another machine by exporting the server configuration so as to satisfy all dependencies while running: apt-get install --download-only <package_name> Can, in effect, apt-get install read a configuration from an exported archive rather than from the local package database? Can the list of packages to be downloaded be retrieved, along with an installation script to perform the installation, instead of the actual packages? (a further level of indirection which would help me schedule this with wget at appropriate times...)

    Read the article

< Previous Page | 205 206 207 208 209 210 211 212 213 214 215 216  | Next Page >