Search Results

Search found 17955 results on 719 pages for 'sub domain'.

Page 543/719 | < Previous Page | 539 540 541 542 543 544 545 546 547 548 549 550  | Next Page >

  • UNC access to TFS SharePoint doesn't work

    - by RobSiklos
    We are using TFS2010 with the SharePoint document portal. We are trying to access the files in SharePoint using UNC paths (e.g. \tfs2010\sites\DefaultCollection\MyProject) and it just plain doesn't work. From my workstation, I actually get different behaviours depending on the path. Case 1: Path = \tfs2010\sites\DefaultCollection\MyProject\ Result: windows explorer reports a network error "Windows cannot access \tfs2010\sites\DefaultCollection\MyProject\" Case 2: Path = \tfs2010.mycompany.com\sites\DefaultCollection\MyProject\ Result: Windows Security dialog pops up asking for my username and password. I tried entering all combinations of my windows username and password (with and without the domain before the username), but no matter what, my credentials are not accepted. I have no problems accessing the SharePoint site through the web portal - it's just UNC which doesn't work. There doesn't seem to be anything relevant in the event viewer on the server. Anyone have any idea what could be the problem?

    Read the article

  • Duplicate of IPSEC certificate template not displaying

    - by Tim Brigham
    I am working on getting IPSEC working on computers outside my domain. I am unable to get a duplicate of the IPSec Offline Request template to show on the web interface. My CA is 2008 R2 Enterprise and the permissions on the template should be adequate to allow me access (my account has read and enroll). If I publish the original template it functions as desired. I also have other custom templates in use (although using autoenrollment). The duplicate is almost identical to the original, excepting that it has a longer validity period and the private key is marked as exportable. How do I debug at this point?

    Read the article

  • PHP and load balancing

    - by StCee
    I have one major domain but the server spec behind it is not good enough. Hence I want to relay the traffic, in particular php-mysql queries to multiple smaller servers. How is that normally be done? (BTW I wonder how much traffic or number of php/mysql request a normal setup on ec2 micro instance can handle? ) I did have a look of EC2 load balancer. But is it only possible to load balance on machines of your own account?

    Read the article

  • Certain websites won't load with DHCP server

    - by Simon E
    Hi, I recently configured a server using Windows Server 2008 to take over from my router using DHCP Server. I disabled my router from assigning IPs too. All works well, except for certain sites that refuse to load e.g. "Firefox could not find www.amazon.co.uk" http://www.google.co.uk/ http://www.amazon.co.uk/ http://www.jobsite.co.uk/ I think it may have something to do with co.uk domain. If I reboot the server everything goes back to normal but of course thats not a good solution to do that every other day. Does anyone know of this problem or what it might be - I am very new to server technology.

    Read the article

  • Hide unauthorized folders from view in webdav

    - by eric.s
    We have a Windows server running webdav. When a Mac connects to it they see the folders they do not have access to as a blacked out folder, and then the folders they do have access to as folders. This does not happen when a client using Linux or Windows connects. All users connect with their domain account credentials. I may not be searching properly to get the information I am looking for, so I turn here. Is this an issue that I should be researching on the Mac side, or ask our network team to be looking into settings on the server?

    Read the article

  • How to find dom0 name from hosted domU

    - by svigan
    I'm actually testing RHEL 5.3 with Xen between two servers in order to have a disaster recovery solution. So I'm playing with moving my domU from one dom0 server to the other server. Unfortunatly when somebody else move the domU I don't have any clue where my domain is hosted. I'm wondering where I can't find my dom0 name with inside domU. I'm looking for something like the gzonename command on a solaris sparse zone. I check inside /proc/xen but I don't see anything special except the dom0 kernel release. Does anybody know a wait to find this ?

    Read the article

  • Getting "is not a valid mailbox" when email sent to primary address

    - by Juan Tarquino
    Anyone outside my domain who sends me an email is getting a bounced email with the following error from Exchange: 550 550 5.1.1 [email protected] is not a valid mailbox (state 14). I can't receive emails from external addresses until I send an email to one of my secondary addresses. Once I sent it to my secondary address, my primary address will start working for a while and fail again after a few hours. We use Exchange 2003. Anyone has any suggestions on what to try?

    Read the article

  • Mixed IP and Name Based Virtual Hosts with nginx

    - by nerkn
    I set up many domains but I dont know how to configure if only ip address is given. say foo.com I have a setup to go web/foo.com/htdocs, I want to 88.99.66.55 ip address like a domain to web/fook.com/htdocs server { listen 80; server_name 85.99.66.55; location / { root /home/web/fook.com/htdocs; } location ~ \.(php|php3|php4|php5)$ { root /home/web/fook.com/htdocs; include fastcgi_params; fastcgi_pass 127.0.0.1:9000; } } resulted [warn]: conflicting server name "85.105.65.219" on 0.0.0.0:80, ignored

    Read the article

  • Setting Up DHCP Server as Failover

    - by Jason Buhagiar
    I have 2 Domain Controllers running at my company; DC1 has the following roles; ADDS DNS DHCP Scope - 192.168.2.100 - 192.168.2.240 DC2 has the following roles; ADDS - Replicated DNS - Secondary DNS I also have 2 PFSense Gateways both have a different ISP Connection, DHCP Relay is not enabled on any of the machines. Can anyone suggest a way for me to install DHCP Server as FAilover on DC2 please, should I use split scope etc... Help is much appreciated :) Thanks.

    Read the article

  • How to share datastores between multiple exchange servers?

    - by Johan
    I have an Exchange 2003 box that is seriously overstressed. I want to transfer its duties to a new and faster box. I don't cannot suffer downtime, so I have to do this stuff live. Here's what I plan to do: Install Exchange 2003 on the new server Set up the new server, so it will accept requests from users for their mailboxes I want to do as little manual set up as possible, because that 'll eat up my time and is too error prone Than I want to transfer my datastores one by one to the new server and have those users (once the datastore in the new server is up and running) to get their data from the new server (without them noticing) I don't have to transfer all the datastores, some of them need to stay on the old box (because I'm still waiting for extra HD space to arrive from the supplier) What steps do I need to follow to do this? The new box has never seen this domain before, the old exchange server is not the DC, we have a dedicated DC.

    Read the article

  • Prioritize file sharing performance in Windows Server 2008

    - by cmbrnt
    I've got a server running Windows Server 2008, and use it mainly for sharing files throughout the domain from a number of disks. It's running on VMware ESXi 4.0, in case that matters. My problem is that when I log in to the server to check user permissions etc, the access speed the files on the remote disks almost grinds to a halt. I havn't been able to measure the speeds, but I would guess it slows down to about 100kB/s as soon as I log in. This is on a gigabit network and the problems are equal for all users, even the ones connected to the same switch as the server. I've assigned 2 GB RAM to the server, and reserved it 1,5Ghz processor power. I don't have to do anything special on the server for this halt to occur. How can I make sure file sharing is prioritized on the server, so no matter what applications I'm using it will always make sure file sharing works properly? Could this be a VMware issue?

    Read the article

  • What is the world wide web? [closed]

    - by think123
    I don't know where to post this question, so please move it if necessary. Ok, so I've heard of how the professional hosting companies can create 'links' to the world wide web to register an unregistered domain. So that's where my question comes from. Is the world wide web a server to which servers link? Is it created by abstract linkage? I'm not sure. Also, what does it mean for the DNS to be updated throughout the whole world?

    Read the article

  • Program for keeping encrypted files.

    - by Giorgi
    I am looking for a program which will encrypt files specified by me and allow me to view/edit/delete those files without creating a virtual disk. I do not want to have virtual disk as a domain administrator can access it so truecrypt is not the possibility. One possibility is to use winrar with password protected archive but winrar serves a different goal so it is not very user friendly for this purpose. If it's possible it would be nice if the program does not creates temp files while I open the files. Any suggestions?

    Read the article

  • One subdomain is not working

    - by BFTrick
    Hello there, My main domain works just fine - www.example.com and a subdomain set up by another developer works as well - sub1.example.com. But when I try to set another subdomain up I go through the process everything seems to work. The software creates the default files where the subdomain files should go. But when I try to browse there it doesn't work. My host uses Plesk to do all of the hosting stuff. What do you think the problem is? I doubt it is some sort of cache issue because I had problems on my phone which I tried after problems on the pc. Maybe for some reason Plesk needs time to set this up? I have used Cpanel before and that works instantly.

    Read the article

  • Windows server 2003SP2 as LDAP replica master for Mac OSX 10.6

    - by FrancoR
    Hello there, we have a single domain controller with Windows 2003 with few child. All the users are in the main DC. We have already created a connection from AD to Mac Xserve 10.6 and can read all the users, but: 1. If the DC goes down (or the net), Mac lose all the users, so no file access, no emails, no nothing. 2. the users are in read only. Mac admin cannot reset password, change attribute and so on. What we need is a stable environment where both AD admins and LDAP admins can manage the users; if one server goes offline the users of the other server should work (email, shared folders) just fine. Thanks in advance P.S. we already tried to connect the MacOSX to Windows LDAP, instead of AD, but we're unable to do it: MacOSX requires DNS IP (gotcha), user admin and password (ok) and a root LDAP password we're unable to find any reference of it in Windows 2003.

    Read the article

  • How does hadoop decide what its nodes hostnames are?

    - by Dan R
    Currently the urls generated by the jobtracker & namenode return either hostnames like bubbles.local or just bubbles. These end up not resolving unless the client machine has specified these in their /etc/hosts file. When I run the hostname command on these machines it returns a hostname complete with the domain (E.G bubbles.example.com) Running a small java test on these machines InetAddress addr = InetAddress.getLocalHost(); byte[] ipAddr = addr.getAddress(); String hostname = addr.getHostName(); System.out.println(hostname); Produces output just like the hostname command. Where else could hadoop be grabbing a hostname to use in its jobtracker / namenode UI? This is occurring in clusters with Hadoop 1.0.3 and 1.0.4-SNAPSHOT from early august. The machines are running CentOS release 5.8 (Final). The generated URLs I'm referring to are like this http://example:50075/browseDirectory.jsp?namenodeInfoPort=50070&dir=/ or http://example.local:50075/browseDirectory.jsp?namenodeInfoPort=50070&dir=/

    Read the article

  • Server 2012r2 VPN DNS

    - by Tyron Gower
    Have an issue where onsite clients cannot resolve VPNusers. but VPN users can resolve onsite machines. example. USER! uses LAPTOP1 USER1 connects to VPN gets internal IP address of 10.243.0.200 USER1 pings SERVER1 - resolve to ip and gets reply USER1 RDP into SERVER1 (inside VPN) USER1 pings LAPTOP1 from SERVER1 resolves to ip address last assigned by DHCP (10.243.0.139) ping fails USER1 pings 10.243.0.200 from SERVER1 gets reply. Running Server 2012r2 It is a domain controller, DNS and VPN server. VPN is just configured with basic default settings. All VPN users have static IP setup in AD. Not sure where to go from here.

    Read the article

  • When does a cached website refresh?

    - by user142485
    If i go to example.com [placeholder for different website] it creates the numerous cache entries in the browser for different items on the page and two entries for example.com. One of which expires in 1.5hrs and the other has 'No expiration time.' What I am wondering is: when does the browser display the cached page and when does it get the newest version from the server (when re-visiting the site)? What do the two different expiration times for the top level domain mean? A web page I went to was temporarily redirecting to a different page. After it was back up, I was still getting redirected to the temp page even though the correct page was up. Would this have eventually resolved itself based on expiration times or does the cache need to be cleared?

    Read the article

  • Setting up Mail (Ubuntu Server 10.04)

    - by Sam
    It seems that mail on my server is functional -- kind of. I have a simple PHP mailer script that is capable of mailing any email address that is not local. e.g [email protected] However, if I try [email protected], it tells me /home/myuser/dead.letter... Saved message in /home/myuser/dead.letter and sometimes You have new mail. What does this mean? Does anyone know what I can do to make my emails actually go through? I'm using google apps for domains -- so basically gmail is hosting my domain's email and I'm not using my own servers for receiving mail. I've set up the nameservers correctly -- I wonder if it's got to do with my sending configuration (what does PHP's mail function normally send with anyway, sendmail?) or maybe it's my receiving configuration -- maybe something makes it act differently for "local" mail

    Read the article

  • Postfix smarthost with diffent relayhosts and sender dependant authentication

    - by mattinsalto
    I've setup postfix as smarthost with different relayhosts and sender dependant authentication. Everything works ok, but I have a performance question. Is it better to send all the email corresponding to a domain through only one account? I mean, now I'm sending each message authenticating to the relay host with the sender credentials. Example: If I have 5 email accounts and I send 10 simultaneous messages from each account, How many times is postfix login to the relay host if I have sender dependant authentication? 5 times? once for each sender 50 times? once for each message If I send all the messages corresponding to one realy host through one account, how many times does postfix login to the relay host? only once? Thanks in advance.

    Read the article

  • How to set up Apache 2 to serve only subdirectories

    - by Lynden Shields
    I have 3 sites which need to be hosted on a web server (apache2 from repo running on Ubuntu 12.04). They are each in their own subdirectory within /var/www/ I would like apache to serve files from the relevant directories only if the directory name is given in the URL, but not serve the /var/www/ directory itself. E.g: http://1.2.3.4/site1/ should work and serve the index from /var/www/site1/index.html, but http://1.2.3.4/ should not serve anything. Currently, I can't get the url to point to the directory. Either I can get http://1.2.3.4/ to serve everything within /var/www/ (including /var/www/site2/secretstuff/), or I can get the root http://1.2.3.4/ to serve one of the subdirectories (/var/www/site1/). This is unacceptable site 1 needs Indexes enabled but the others must not. I just want to make site1's config only respond to requests of the form http://1.2.3.4/site1/* and not handle requests of the form http://1.2.3.4/ I do not have a domain name set up so I can't use subdomains.

    Read the article

  • Subdomain on a separate server (Windows/IIS) won't load default web site?

    - by DOTang
    I've got a domain hosted under godaddy.com. I set it up so that subdomain.mysite.com points to a different server which uses Windows/IIS (say for example ip 1.1.1.1). When I go to the subdomain I get no errors, but no webpage loads. In IIS under the default website I added a bindings for subdomain.mysite.com with the ip of 1.1.1.1, but still nothing loads, just a totally blank page. I know for sure the subdomain host is working correctly because when I ping my subdomain, the correct IP shows. What am I missing to get this working?

    Read the article

  • "No input file specified" - unable to access phpMyAdmin using debian squeeze

    - by guiltybyintent
    I have installed phpMyAdmin on my VPS LAMP server (Debian Squeeze/Apache2/MySQL/PHP5), but am unable to access it: //my-ip/phpmyadmin/ and //my-domain/phpmyadmin/ both produce the following error message: "No input file specified". The phpMyAdmin FAQ identifies this as a permission problem, but the suggested solution seems not to apply to my situation. Every other solution I have come across involves removing/purging and reinstalling phpmyadmin - which I have done several times, always to the same result. Previous posts in this forum typically relate to Nginx, which I have not installed. Thanks in advance for any help!

    Read the article

  • ServerA can't access ServerB through UNC path. How do I troubleshoot?

    - by cparker4486
    All my servers are in the same domain on the same network. ServerA cannot access ServerB through UNC path. However DNS resolves correctly for ServerB from ServerA. For example "ping ServerB" works as expected. ServerB can reach ServerA through UNC path without any problem. Likewise all other servers and clients can reach ServerB through UNC path without any problem. The software firewall on ServerB is not turned on and there are no hardware firewalls getting in the way. This makes no sense and I don't know how to troubleshoot it. Nothing shows up in the event logs of ServerA and ServerB has one event (repeatedly): Event ID: 3019, Source: MrxSmb. ServerA is Server 2008 R2. ServerB is Server 2003 R2. I have other 2008 R2 servers that connect without any problem.

    Read the article

  • Vhost in Apache only working locally?

    - by Gasman
    Ok, I have added lines like: 127.0.0.1 somedomain.com Or some other domain that points to my routers IP, and is forwarded, but I get to the main site, but I want it to go to the subfolder I defined in my httpd-vhosts.conf: NameVirtualHost somedomain.com:80 <VirtualHost somedomain.com:80> DocumentRoot "D:/Apps/xampp/htdocs/somedomain" ServerName somedomain.com ServerAlias somedomain.com </VirtualHost> So, locally somedomain.com works, just remotely it goes to the root htdocs. So If I use a *:80 wildcard I works, but then everything points to the subfolder and all the other vhosts seem to get ignored. Any Idea why this is?

    Read the article

< Previous Page | 539 540 541 542 543 544 545 546 547 548 549 550  | Next Page >