Search Results

Search found 12211 results on 489 pages for 'industry standard'.

Page 246/489 | < Previous Page | 242 243 244 245 246 247 248 249 250 251 252 253  | Next Page >

  • What Windows licenses are required to run additional terminal service sessions

    - by John P
    We need to build out a server running Windows 2008 R2 Standard that can allow up to 10 simultaneous RDP/Terminal Services connections and I'm a little confused about how the CAL licenses work. From one source I was told I needed 10 "server CALs" and an additional 10 "RDP CALs" (total of 20 CALs). From another, I was told I just needed the 10 "RDP CALs", which implicitly came with the server CAL. The Microsoft licensing website (http://www.microsoft.com/windowsserver2008/en/us/licensing-rds.aspx) seems to support scenario #1, but it is still not real clear what those server CALs are needed for. Also, can we use the 2 "built-in" RDP clients, meaning we only need to purchase 8 CALs to reach a total of 10?

    Read the article

  • Why Standards Place Limits on Data Transfer Rates?

    - by Mehrdad
    This is a rather general question about hardware and standards in general: Why do they place limits on data transfer rates, and disallow manufacturers from exceeding those rates? (E.g. 100 Mbit/s for Wireless G, 150 Mbit/s for Wireless N, ...) Why not allow for some sort of handshake protocol, whereby the devices being connected agree upon the maximum throughput that they each support, and use that value instead? Why does there have to be a hard-coded limit, which would require a new standard for every single improvement to a data rate?

    Read the article

  • How to migrate an SQLServer 2000 database from one machine to another

    - by Saiyine
    This January I'm migrating our main SQLServer 2000 based database to a beefier server. Is there any standard procedure or documentation on how to do it? I need to replicate all at the new server (databases, jobs, DTSs, vinculated servers, etc). Edit: I mean SQLServer 2000 on both ends! Edit: Be calm, people, I just crossed the versions from another software I posted about at the same time as this. Effectively, I even checked the wikipedia to be sure version 8 was 2000. Don't need to flame that much about what is just an errata.

    Read the article

  • Why shibboleth IdP idp-metadata.xml recommends 8443 for SOAP?

    - by toma
    After the install.sh of 2.4.0 Shibboleth Identity Server, the idp-metadata.xml file is created. Why is that? Is not enough secure to use the standard HTTPS/443 port? <ArtifactResolutionService Binding="urn:oasis:names:tc:SAML:1.0:bindings:SOAP-binding" Location="https://idp.example.com:8443/idp/profile/SAML1/SOAP/ArtifactResolution" index="1"/> <ArtifactResolutionService Binding="urn:oasis:names:tc:SAML:2.0:bindings:SOAP" Location="https://idp.example.com:8443/idp/profile/SAML2/SOAP/ArtifactResolution" index="2"/> <SingleLogoutService Binding="urn:oasis:names:tc:SAML:2.0:bindings:SOAP" Location="https://idp.example.com:8443/idp/profile/SAML2/SOAP/SLO" /> <AttributeService Binding="urn:oasis:names:tc:SAML:1.0:bindings:SOAP-binding" Location="https://idp.example.com:8443/idp/profile/SAML1/SOAP/AttributeQuery"/> <AttributeService Binding="urn:oasis:names:tc:SAML:2.0:bindings:SOAP" Location="https://idp.example.com:8443/idp/profile/SAML2/SOAP/AttributeQuery"/> Thanks, Tamas

    Read the article

  • How do I set zone priority in Microsoft DNS?

    - by Justin
    I have a standard small network setup (20 users) on Active Directory. All Windows machines have a primary DNS server as the AD and a secondary DNS server as Google PDNS. I want to setup a DNS entry that exists in real DNS but set it up on our DC so that local requests would route this public domain to a local development machine on the network. I setup the zone in DNS which results in the clients resolving the public FQDN to our internal IP. However, sometimes it still resolves to the "real" value (I check by pinging it). Is there some way to give the zone definition in my DC DNS higher priority? Or will the client that has secondary public DNS always at sometimes have a competing entry for this zone?

    Read the article

  • Is defragging right for me?

    - by blade
    Hi, I am using Hyper-V on my Windows Server 2008 R2 DC x64 machine. I am also using standard SATA drives. I read some threads on here about defraging but could not reach a conclusion about whether or not I should use defragging. Can anyone shed some light on whether this will be right for me? Furthermore, what tool is best? There seems to be 3: http://www.perfectdisk.com/products/business-perfectdisk11-server/key-features http://www.diskeeper.com/diskeeper/home/server-edition.aspx?id=40279&wid=7 http://www.perfectdisk.com/products/business-perfectdisk11-hyper-v/learn-more Anyone have experience with this?

    Read the article

  • Blocked connections passing through firewall. What is wrong?

    - by Kiranu
    In our company we have a small business router (Cisco RV082) on which we are using its standard configuration (block all incoming traffic). We also have an SMTP relay configured (using WS2008R2) so that our internal applications can send email through google apps (which requires authentication). The thing is that the server was being used to send spam. We fixed the problem by only allowing the server to relay email from our internal IP address range (10.0.0.0/16). My concern is that there was a way by which external IPs connected to the network and that underlying problem has not been fixed, but I cannot imagine how these machines connected. Any thoughts?

    Read the article

  • Adding a Windows Server 2012 Essentials server to an existing domain, without migrating the AD

    - by TiernanO
    I have an existing Active Directory in house, a mix between a Win2K8R2 and Win2K3 domain, and i would like to test out Windows Server 2012 Essentials BETA on the network. When walking though the install, it gives me the option of a new domain, or migrating from an existing domain. when clicking existing, it tells me i can only have one SBS server running on a domain at a time... So, i dont have any existing SBS servers in house (both are full standard or enterprise editions) but i do plan on keeping at least one of these extra servers running... So, how do i get a 2012 Essentials server to join a domain, and not migrate the existing domain? or if i do migrate, can i still get one of the other boxes to act as secondary controllers?

    Read the article

  • Windows VPN not authenticating from ADSL to Wireless link

    - by deanvz
    I have a normal windows VPN on a computer connecting to a 196.201.x.x/24 IP. If this VPN tries to connect from any address in the 41.x.x.x range it cant get there. The server is a normal windows 2008 server, running exchange with a PPPoE IP natted to a public IP on the public gateway of the wireless network as the server is on site and its connectivity is derived from Mikrotik RB's. The computer on the 41 range can traceroute and ping the server, but the VPN does not authenticate. When on the network or any other, the VPN works fine. Is there something that could be configured on the VPN client? All firewall settings of a standard ADSL router have been checked and found to block only ICMP's. Is this a VPN configuration problem or a network issue?

    Read the article

  • Can a power failure or forceful shutdown damage hardware?

    - by Vilx-
    Can computer hardware suffer damage from forceful shutdowns (holding the power button for five (5) seconds) or power failures? I believe that normal PC hardware does not suffer from this - after all, it's not much different than what they experience under a standard shutdown. But elsewhere I've read tht another person thought that it could do physical harm to the hard drive and possibly other components as well. He also said that the journaling features of filesystems are useless in face of power failures and were intended to help mitigate damage from system crashes. I think this is nonsense, but then again I lack the experience and knowledge to say it with certainty.

    Read the article

  • Screen goes grey in Remote Desktop Connection from windows 7 to ubuntu server

    - by user1001328
    I recently got a 64 bit ubuntu server based on Ubuntu Desktop 12.04 "Precise Pangolin". However I can't seem to connect to it properly from my windows 7 64 bit machine. I can get to the login screen for ubuntu, and it seems to succeed in logging in, but then the screen just goes grey before anything comes up. (too low rep to post a picture) I installed xrdp on the server and attempt to connect using the standard windows Remote Desktop Connection program. I also ran the following 2 commands on the ubuntu machine to try get it to run but it was no help. echo "gnome-session --session=ubuntu-2d" > ~/.xsession and sudo apt-get install gnome-session-fallback

    Read the article

  • 2.5" SSD in a 3.5" desktop drive bay normal?

    - by Cory Petosky
    I'm a little out of touch with modern desktop builds, but I recently built a new machine anyway. I want to pick up a SSD for my I/O intensive apps (EverQuest 2 and Flash CS4 mostly), but I'm having trouble finding any in the 3.5" form factor. Is this expected, or do I fail at product search? And, if it is expected, do 2.5" SSDs come with an adapter to fit them in my desktop machine, or will I need to purchase one separately? And, finally, if I do need to purchase an adapter, are they pretty standard or are some better than others. General advice is most welcomed, though specific product recommendations in addition would be helpful!

    Read the article

  • OS X Terminal using default ANSI colours for every theme

    - by FrogBot
    For some reason, every theme I palette I've had installed for the OS X Terminal.app has reverted to using the default ANSI colours. The themes have been working fine up until this point and I can't seem to determine what could have caused them to revert in this way. For reference, my standard Solarized Dark colour scheme should look like this … … but it currently looks like this: A quick look in the preferences panel shows that all the colours are correct save for the ANSI colours which have reverted to their defaults. I don't know what other information would be helpful but if you need any other info to help me troubleshoot just ask and I'll update as quickly as I can.

    Read the article

  • How do I mount an Exchange mail store from the Windows Command Line?

    - by Cypher
    Our Exchange server is running Exchange Server 2003 Standard on the Windows Server 2003 platform. We're dealing with the mail store size issue, where if the mail store goes over the limit, it gets dismounted. While we are working with the powers-that-be on a policy that will prevent this happening in the future, I would like to see if it is possible to re-mount the mail store via the Windows CLI. I'm already monitoring the Event Logs and alerting on mail store warnings and dismounts - I'm just tired of getting up at 5am to manually re-mount the store while the political wars ensue. My alerting tools have the ability to execute a batch script when an alert is generated. I would greatly prefer a native CLI option. I'm not too keen on running some random vbscript found on the Internet and I don't really care to spend my time debugging someone else's code. PowerShell might be an option, if it can be triggered from the CLI.

    Read the article

  • low speed web application, Server problem or Application

    - by Ashian
    Hi, I have a web application written by asp.net (c#) sql server 2005. we host it on 2 dedicated server ( IIS and SQL server ) From some month ago , in some days of week we have many reports about speed issue. we have some other application on this server using same database. when we have speed problem all aplication on these server have this problem, but applications on other server in same data center work correctly. ram and cpu usage are ok. how can I check that the problem related to internet connection or my application design? which parameters must be checked. Some other information In applications users can upload several files to server , each file up to 3 MB. we use a sql web admin application, on same server that has same problem, this is a standard application which work perfectly on other servers. Thanks

    Read the article

  • How can I remove OLD history from Google Chrome?

    - by Norman Ramsey
    I'm working on a laptop with a modest hard drive, and 500MB is taken up with Google Chrome "History Index" and "Thumbnails" files. Some of these files are a year old. Chrome offers me the option to remove recent history, but I want the opposite: I want to remove old history. (Ideally I would remove the least recently used history information, but I don't expect to be able to do that.) Anyone have any ideas? I'm running the standard Debian google-chrome-beta package.

    Read the article

  • ldirectord refusing connection when nginx redirects from http to https

    - by Adam
    I am running ldirector as a load balancer to an nginx front end server. If I setup a redirect from http to https and connect directly to the nginx server, all is well. Connecting via ldirector causes my connection to be refused. I can connect normally via http or https through ldirector when I don't have the redirect in place. To add to my confusion, if my application issues a redirect from http to https, it works. I am testing this via curl on the command line. (curl: (7) couldn't connect to host vs a response) I am using the standard ldirectord config (http://www.ultramonkey.org/3/topologies/config/lb/non-fwmark/linux-director/ldirectord.cf) the http and https parts. My nginx config for the redirect is simply: location / { rewrite ^(.*) https://$host$1 permanent; }

    Read the article

  • Server memory issues, and expected level of service from hosting company

    - by Greg
    I'm involved in maintaining an Ubuntu VPS which runs our django websites (nginx/apache/mod_wsgi) and we've been having some memory spikes which have either caused the database to die, or induced kernel panic when the memory management system can't find any killable processes. I'm working on fixing the memory spikes, but I'm wondering whether there's anything I can do to better deal with the problem if it occurs again. Are there any tools I could use to detect the memory spikes and then, say, kill the offending process and email the server admin to fix it up? Killing off one website so that the server can remain operational is certainly preferable to the whole thing falling over. Also, we were charged $600 for after-hours service because we had to get the hosting company to restart the server - is this standard practice among hosting companies? Another provider I work with provides a panel with which I can stop and start the server myself, and given that a restart was all that was needed, $600 seems mightily excessive. (That's NZD, it's around $445 USD)

    Read the article

  • Apple Software Update Server for Windows

    - by Matthew Iselin
    We have just added a few iMacs to our system and we've found that they all want to download about 1.6 GB of updates... not so great when we only have limited monthly bandwidth! All of our Windows machines just use WSUS, which works great in our environment. It'd be nice if we could do the same for the iMacs without purchasing an additional Mac for a server role. So is it possible to run Apple's Software Update Server on a Windows server? Or do we need to look at purchasing a Mac in order to distribute updates across our clients? Alternatively, could we set up one of the iMacs to run the update server for the other iMacs whilst it is being used as a standard machine (ie, not installing OSX server, and keeping it available for staff to use)?

    Read the article

  • help building a PC that can image a dozen hard drives simultaneously

    - by Bigbio2002
    Not sure if this belongs on here or SuperUser, but here goes... I'm trying to figure out how to make a mass hard drive imaging PC out of COTS parts. A dedicated imaging device can do 10 drives at a time, but costs several thousand dollars. So far, I'm thinking to use several 3-port PCI-E Firewire cards, and use some kind of Firewire-to-IDE adapter to connect the drives themselves. The "software" would consist of scripting diskpart, or some other imaging utility. The problem is that I can't seem to find any sort of adapter. I could use standard external hard drive bays, but then I'd have a dozen power cables that I need to plug in. Ugly, messy, and inefficient. I picked Firewire over USB not only for better transfer speeds, but also because FW can deliver power over the bus (and could theoretically power a hard drive). Does anyone have any input on this?

    Read the article

  • DICOM Image viewer for Windows

    - by Alin Purcaru
    In case you were wondering DICOM (Digital Imaging and Communications in Medicine) is a set of standards in medical imaging. I have a CT result saved conforming to that standard and I want to be able to see the scans. The directory structure looks like this: 123456 (folder) |- 2342345 (file) |- 2342346 (file) |- 2342347 (file) |- [...] (more files) DICOMDIR (file) A Google search showed a lot of results for DICOM viewer and I even tried one but it only shows the thumbnails of the images. Could anyone recommend a good viewer so I don't have to try them all? Maybe with export functionality. Thank you, Alin

    Read the article

  • Reimage several of the same model PC by copying the hard disk

    - by Tim Lehner
    I have several of the same machine (Dell Precision T3500) that originally came with Windows 7 Pro (there is a Windows 7 Pro OA product key sticker on top of each one). These were initially downgraded to XP by a former IT colleague, but are now back in the pool of unused machines. I am looking to format and install Windows 7 Pro on one of them using the latest official Win7 installer ISO with SP1 and then clone the hard drive to the other machines. The questions then, are: Does this even make sense, or is there a better way to do this considering I'm only doing it to 6 machines? Does this jive with MS licensing (we are the original purchaser of the machines, and thus, presumably, the license holder...but I'm no expert)? Is it possible to apply the individual 25-char product keys to the cloned machines after cloning? How far can I go in the config process (installing corporate standard apps and such) before I have to clone to the other machines considering the licensing/key questions?

    Read the article

  • What if a computer with Ethernet

    - by George Nixon
    I'm just revising for an exam on Networks and Data Communications, and there's one thing I don't get about CSMA/CD and Ethernet. It's supposed to be fairly stable, for instance if a computer drops out of the network, it's not a problem like it might be in a token ring network (I think). But Ethernet works by all the other computers waiting for the currently transmitting computer to finish what it's doing, and then the others use CMSA/CD to determine who goes next. What if one computer malfunctioned and kept sending a continuous stream of data in an infinite loop? In fact, is there a standard time for pcs to transmit before they yield to others?

    Read the article

  • Ubuntu on an XPS 14 Ultrabook with mSATA cache and 500GB HD - how to partition for dual boot?

    - by JDS
    I am getting an XPS 14 ( http://www.dell.com/us/p/xps-14-l421x/pd ) and I want to dual-boot Windows and Ubuntu. This thing has a 500GB standard HD and a 32GB mSATA that can be used as cache. Does anyone know how this thing is partitioned? Is the OS installed on the mSATA drive and data is on the big HD? Is there a BIOS controller or maybe even a Windows driver that makes the mSATA drive and 500GB HD appear contiguous? I get the impression that something makes the mSATA be used invisibly as cache, but I can't find any technical documentation how that works. My primary concern here is wrt dual-booting Ubuntu. I want to know if I need to partition the mSATA separately, or the big HD, or just partition the "magic" contiguous disk space that appears available to the OS.

    Read the article

  • SSL Certificates, two-way authentication and loadbalancers

    - by 5arx
    We're looking to implement two-way authentication with client certificates for a privileged subset of our application users. The idea will be that if a certificate is detected the user will be asked for an additional password/PIN and that will be used to verify the certificate and user. Ordinary users will continue to authenticate themselves via the standard login mechanism. Our production environment (hosted by a well-known company) comprises load-balanced application servers and I'm unclear as to how this set-up will handle the certificates and I'm not certain if there are any pitfalls I should be aware of. I would very appreciate some thoughts, comments or real-world advice on the subject.

    Read the article

< Previous Page | 242 243 244 245 246 247 248 249 250 251 252 253  | Next Page >