Search Results

Search found 3246 results on 130 pages for 'planning'.

Page 66/130 | < Previous Page | 62 63 64 65 66 67 68 69 70 71 72 73  | Next Page >

  • What are the components required to run taskpads on a workstation?

    - by Darktux
    we are planning to implement task pads in our enviroment for delegation for user administration.I have few questions regarding this. 1.) To run a taskpad (with AD users and computers) ; does the workstation contain whole set of Administrative tools oris there a way just to compy dsa.msc to all workstations and get done with? Note: All tak pads will be shared on a file share and users access them via powershell scripts. 2.) We are creating MMC's in Windows 7 and keeeping them in a share; will the work with Windows XP too or do we need to develop 2 versions of them? We are aiming to keep away software from workstations and maintain it as much as possible on centralized file shares. Please shoot me any questions or clarifications pertaining to my query.

    Read the article

  • Is it safe to expect Sun Java 6 to supported for the life of RHEL 6?

    - by Ophidian
    I'm in the planning stages of a java application that we're targeting for Red Hat Enterprise Linux 6. Unfortunately, we're stuck at RHEL 6.1 which does not ship the java-1.7.0-oracle package set (they were added in 6.3) and I don't really have any control over when we will be upgraded to the more recent version. I don't have any specific technical requirements to use Java 7, but Java 6 is going to hit public EOL in February 2013. Am I safe to assume that since Red Hat (and subsequently Oracle with its Oracle Unbreakable Linux) has shipped a copy of Java 6 in the java-1.6.0-sun package, it will support it for the entire 10 year support life of RHEL6?

    Read the article

  • server performance metrics report and practicality

    - by Anjesh
    I have a need of preparing web server (apache-php) performance report containing important metrics like CPU usage, disk io, memory usage on user basis. Couple of domains are hosted in the same server and they run from separate users using fcgi. The reason being sometimes some hosted applications take lots of cpu usage, making the server slow for other applications (running as separate users). i am planning to develop scripts for this, as i can't seem to find any simple utilities for this purpose. This script will take snapshots of the user wise metrics at defined periods say 15 minutes and record it. Any abnormalities will be reported via emails. How practical is that? also would be interesting to know what else need to be recorded.

    Read the article

  • Sharing Internet over Wireless Router

    - by Alandt
    I have a very strange question today - how do I share my dial up internet (yeah I know you are gonna say that is slow but broadband internet isn't available in my area), so dial up and 3G connection is all I got. I also have a Vodafone USB 3G modem that picks up 3G network, I am planning to use my Vodafone 3G modem in the day since I have free dial up internet from 7:00 pm in the night untill 7am the next morning. Some additional details: * My PC is running Windows XP Professional SP3 * I have a Sitecom Wireless Router 150N X1 WLR- 1000 I would appreciate it if anyone can provide me with a step-by-step guide! Thanks

    Read the article

  • VM: Windows 7 host, Linux guest, VT-d?

    - by chx
    I am sick of the driver issues of Linux. So I am planning to switch to Windows 7 as a host and virtualize my Linux into it. My laptop has integrated Intel graphics and supports VT-d. For speed reasons I would like to assign that card to Linux. Now, Parallels could do it but this page says Note: If you have only one PCIe video adapter, its name will be grayed out in the PCI Devices list and you will not be able to allocate it to your virtual machines. I would be perfectly fine with a headless Windows 7 (I can remotely admin from other computers or just the Linux guest) -- is there any VM software that doesn't have this restriction?

    Read the article

  • How can I force all requests to be SSL when using EC2 load balancer?

    - by chris
    I currently have a single EC2 instance which is forcing all requests to be secure by using mod_rewrite: RewriteEngine On RewriteCond %{SERVER_PORT} !443 RewriteRule ^(.*)$ https://%{HTTP_HOST}$1 [R,L] I am planning on moving to a load balanced setup, with multiple back-end instances. If I set up my EC2 load balancer with my certs, do I need to use SSL to communicate between the LB and my instances? If not, is it as simple as replacing the RewriteCond with RewriteCond %{HTTP:X-Forwarded_Proto} ^http$ Edit: I tried using the x-forwarded-proto, but it does not appear to work. Is there another way to detect if someone is connected to the LB via SSL?

    Read the article

  • Steps to install solely ubuntu 13.04 on Dell inspiron 14z ultrabook with SSD+HDD

    - by rishy
    I have tried a few things like disabling the Intel smart response, choosing AHCI in BIOS. But there are certain problems I am still facing. I can't see my SSD during the installation of ubuntu (I am planning to install Ubuntu on my SSD and other files on HDD). When I run Ubuntu my laptop gets overheated and battery backup reduces to 90 minutes. (I guess it's related to my graphic driver ATI Raedon HD 7570). Cooling fan seems to run at its fullest, it was working much better in windows. So, overall I wanted to know what are the exact steps I need to follow to install Ubuntu on my SSD and then use my HDD to keep other files, How can I get rid of overheating and battery backup problem?

    Read the article

  • upgrade to windows8

    - by Siddharth Warrier
    I have a dell inspiron N5010 laptop. I am planning to upgrade to windows 8 and the existing version is windows 7. So when I planned to upgrade I found that in the website of dell they have not updated the drivers for windows 8. I had installed a preview version of windows 8 and installed the drivers of windows 7 so it got installed and is working fine. So shall I upgrade it to windows 8 and use windows 7 drivers???? or wait till the drivers are updated??????

    Read the article

  • Do I need to reserve space before installing Ubuntu alongside Windows 7?

    - by CRM Junkie
    I had Windows 7 32 bit on my existing system, but I am planning to install Ubuntu alongside it. So, I just decided to do a fresh installation of both the operating systems. When I insert the Windows 7 DVD, I can create 3 partitions at maximum, with one being the one where Windows 7 will be installed. I just wanted to know – do I need to keep some un-partitioned space for Ubuntu to install? By "unpartitioned space" I mean the space left after creating 3 partitions for Windows 7. I have a 500 GB HDD, so the three partitions I would be creating are 120 GB, 120 GB and 120 GB. The rest is shown as some logical drive, is that unpartitioned space? Can I install Ubuntu over there? I am pretty sure the space shown as logical won't be available as drives when I log into Windows 7. Is that space lost or can I use that to install ubuntu?

    Read the article

  • Returning Shapes to Default in Excel

    - by Craig
    Hi, I have around 70 shapes in a planning document i use for work, everything is fine but i am trying to add a new feature. These shapes are changed using edit points each week to show up on a map, but sometimes shape "A" may not get used in which i just want to turn it back to a default size along with all the other shapes. Does anyone know how i could achieve this via a Macro, i have tried lots of things and searched everywhere but i am at my wits end... If a shape is not default, set all non default shapes to default size. Thanks in advance

    Read the article

  • Does NTFS performance degrade significantly in volumes larger than five or six TB?

    - by Josh Yeager
    One of my customers is planning to set up a new document store, which will probably grow by 1-2TB per year. One of my co-workers says that Windows performance is extremely bad if it has a single NTFS volume that is bigger than five or six TB. He thinks that we need to set up their system with multiple volumes so that no single volume will exceed that limit. Is this a real problem? Does Windows or NTFS slow down when the volume size reaches several terabytes? Or is it possible to create a single volume of 10 or more TB?

    Read the article

  • Triggering GDM login on a remote machine

    - by creator
    I have to briefly describe the situation. We are planning to make a computer classroom with workstations running Ubuntu 10.04. Since making accounts for each student has not been considered reasonable, we decided to make accounts for each student group. We don't want students to share their passwords between groups so the solution would be not to give them passwords at all, but let the teacher log them in instead. Obviously he shouldn't go from one machine to another typing in credentials by hand, so we need some script that will connect to a remote machine by ssh and make GDM (or probably any other login manager if GDM cannot serve this purpose) log in specified user. I couldn't find any solutions, as well as I haven't noticed anybody in similar situation asking for help, so my question will be: can the scheme described be realized and if yes, then how? Thanks in advance.

    Read the article

  • How do I backup a git repo?

    - by acidzombie24
    I am planning to switch from SVN to git. With svn I just copy my repo folder when I want to back it up. However git doesn't have one so what do I do? Should I create a clone on a separate drive and update by pulling from my project? Then I can burn/archive this folder and it will have all the history? This is probably obvious but I want to make sure when it comes to backups. I still pretend there is a root repository.

    Read the article

  • Workaround for API limits [closed]

    - by blunders
    Problem: Planning on building out a client services company that requires access to APIs. Most APIs are limited based on user, IP, etc. - and even though the API calls would be on a per client basis, there's no way to get usage not tied to IPs. (Theoretical) Solution: Have each client install on their network a proxy/VPN that would allow my systems to connect and use their assigned usage. So, it's possible there's a better solution than the one I've thought of, but it's the only one I've been able to come up with.

    Read the article

  • How would I measure the amount of RAM needed per Glassfish domain? [closed]

    - by oligofren
    Possible Duplicate: Can you help me with my capacity planning? In our test environment we have a lot of apps spread out over a few servers and Glassfish domains. To make versioning easier I would have liked to have one Glassfish domain per customer per app (kind of like a heavyweight version of lots of jetty instances). But I have heard that Glassfish is kind of heavy on the resources, and so I would need to measure approximately how many instances would fit in the available RAM. These are low-traffic/low load testing servers, so CPU is not really an issue, though RAM might be. How would I get an approximate measure of how much RAM is needed? This is one Glassfish 3 instance with one heavy EAR application deployed. top? jvmstats? ??

    Read the article

  • WLC 4400. Get IP

    - by Edward Connelly
    I have three SSID. one is a guest network and I'm having problems with lease times. Its set for web authentication so it passes IPs out even if someone doesn't try and authenticate. Is there a way to get that IP back and not hand another to that mac for a period of time? Poor planning and we didn't allocate enough address space, and we are sandwitched. We could go with another address, but we would have to reconfig lots. since its guest its vlan'd with no access to anthing but the firewalls. would have to change ACLs and routes and interfaces. if it isn't possible we will just have to go that route, just thought there might be an easier solution. session timeout is enabled with 1 hr lease time is set to 4 hrs. no helper DHCP and internal DHCP is set with the space split across three WLCs (60 waps split across those as well).

    Read the article

  • OpenVPN: Single certificate authority, multiple VPNs

    - by darwish
    The company in which I work has a single site (I'll refer it as "Site A"). There are several private networks within site A. We have a running instance of OpenVPN which allows some employees to connect to one of the private networks in site A. We're planning to extend our facilities to another site (which I'll refer as "Site B") and we wish to connect both sites using OpenVPN. The VPN which will connect sites A to B will be a trunk link, meaning it will have access to all networks. If we use the same certificate authority for both VPN servers, this will allow the employees, which can only to one of the private networks within site A, to connect to the site-to-site link, which will give them access to all networks. Off course this is undesirable. Using 2 different certificate authorities seems like the obvious solution, but it doesn't feel right. I wounder if there's a way to maintain permission control within a single certificate authority.

    Read the article

  • How to import Exchange calendars in Google Calendar (No Desktop Outlook)

    - by Tonny Madsen
    I would like to import my calendar from my company Exchange 2010 server as a new calendar in my private Google account. The primary use is to easy my planning when at home with my laptop. I already have access to both the Google and Exchange calendars on my Android phone, but the size of the screen makes it rather difficult to plan anything serious. Note that I run OSX on my laptop I would prefer a Internet based solution to any software on my own laptop. I do not want an Outlook add-in I do not want to include the work related events in my main calendar I do not want a mobile solution I can pay for the solution

    Read the article

  • Hooking up many different external HDs simultaneously

    - by cbizz
    I need a large amount of external storage for an upcoming project. I'm planning on purchasing 10 2TB external drives. I need them all hooked up to a single machine at the same time. What issues will I run into? I plan on using 2 power strips and having them all externally powered from the wall. I will use a USB hub to plug in all the drives. I need drive access time to be as fast as possible. I am using Ubuntu Linux(64 bit). Will I be able to mount 10 drives?

    Read the article

  • ZFS Configuration advice

    - by rbarrette
    I need some advice on configuring ZFS. Here is what I have: Physical Disks: 4x 3 TB 2x 2 TB 2x 1 TB What is the best configuration for my Vdevs and storage pool. I want to maximaze space but still maintain redundancy. Should I just get 2 more 3TB's and just create 2x 3-3TB raid2z storage pools? Create a 1x 4-3TB raidz2 vdev? Can I put redundancy at the pool level and create individual vdevs for each drive and then add 2x 1TB+2TB striped vdevs to keep all vdevs the same size. Keep in mind I do need to migrate data from the smaller drives and am planning on adding more 3tb drives later on. What do you think?

    Read the article

  • How can I make a non-destructive copy of a (NTFS) partition?

    - by violet313
    I want to recover some deleted files from a healthy NTFS partition on an undamaged hard-disk. In order to leave the partition undisturbed, i plan to use dd to clone the partition to a raw image file & then attempt recovery from that mounted clone. Will dd if=/dev/sd<xn> of=/path/to/output.img perform a non-destructive copy ? Is attempting a restore from a clone using dd the best approach? [edit, wrt Deltiks answer, i need to be a bit clearer about what i'm asking] eg: are there some s/w that can do something more with the original sectors ? eg: if it was a damaged hard-disk i am aware that any kind of read is potentially destructive. but assuming my disk head is not going to suddenly spaz out etc, am i reducing my chances of a successful recovery (at any cost) by using an apparently non-destructive single read of my undamaged hard-disk. (btw: i am planning on using ntfsundelete & testdisk for recovery)

    Read the article

  • Windows Vista to Windows7 upgrade

    - by Kalls
    I have a Windows Vista Home Premium and I tried installing SP2 on it and It prompted me to install SP1. Then it stopped working when I installed SP1. I don't have back up that I can use to restore it. When I use the Windows Repair mode I am unable to get access to few options none of which is working. I do have access to the command prompt. Is there way to restore this? If I am planning to upgrade to Windows 7 Home Premium what are the steps?

    Read the article

  • Building a network at home, what cables to use (if any)?

    - by Faruz
    My house is currently in ruins and am building it. While doing so, I wanted to design a home network. My main objectives are surfing and HD streaming. The house is one-level, 100 sq/m (about 300 sq/ft), and one of the rooms is a safety room with Reinforced concrete walls. About a year ago, when I started planning, I thought about putting Cat 6 STP cables in the walls and create network points in the rooms. Should I use STP or FTP? I heard that STP is a problem regarding connectors and stuff. Is it really beneficial? Will it work OK if I transfer the wire together with the telephone line? Should I maybe go with WLan and count on 802.11n to enable me to stream HD across the house? is 802.11n that good?

    Read the article

  • What's the best way to be able to reimage windows computers?

    - by mos
    I've got a low-end machine for testing our software. It needs to be tested under various versions of Windows, so I was planning installing each one on its own partition. Then I realized that after testing our software, I'd want to roll back to the previous, clean state. I don't want to use any virtualization software because it tends to interfere with the workings of our app. That said, what's the best way to achieve my goal? Norton Ghost? Edit: I work for a pretty monstrously huge organization. Money is no object here (and sometimes, if the wrong people get wind of it, "open source" software is bad).

    Read the article

  • Ubuntu+Mono+Postgres+ASP.NET 4.0. No problem?

    - by wreck_of_u
    Would this be ok? I'm an ASP.NET developer and I'm planning to build "portable" web app servers based on Atom D510 mini-ITX. I have ran Ubuntu 10 with MySQL along with a separate IIS machines (win 2k3, 2k8) before with no problems. But now I'm thinking of "packaging" a web/db server into one small, cheap machine. I thought of Ubuntu/Mono/Postgres/ASP.NET, that it would be a good idea but I'm not sure? I have not actually tried it yet. Your thoughts?

    Read the article

< Previous Page | 62 63 64 65 66 67 68 69 70 71 72 73  | Next Page >