Search Results

Search found 9091 results on 364 pages for 'thinking'.

Page 129/364 | < Previous Page | 125 126 127 128 129 130 131 132 133 134 135 136  | Next Page >

  • what are security implications of running Ubuntu inside Windows 7?

    - by EndangeringSpecies
    I am thinking of switching to Ubuntu as a way of making web browsing more secure. So, suppose I will go the easy route and run Ubuntu as an app inside Windows and then run Firefox inside of that. What will this do to the security given the current threat environment? E.g. do most online threats nowadays target the browser and flash (which presumably would be safely sandboxed inside easy to wipe Ubuntu environment) or do they target the Windows TCP-IP stack where Ubuntu would give no protection? Well, most likely the above question does not come near to covering all the security implications of this setup :-), so please do discuss whatever other issues that may be relevant here.

    Read the article

  • Win2008 - restrict VPN user permissions

    - by Sebas
    Windows 2008 R2 SP1 Foundations file server with no AD, only workgroup sharing some folders, and now a RRAS server. Shared folders are open to everyone in the office (XPs and Sevens) without accounts/passwords, but I was thinking about partially limiting access to the new "VPNuser" account. I'm new to Windows Server and its permissions settings: I thought about denying access to vpnuser through NTFS rights in some folders. It doesn't work, but now I'm guessing that the vpnuser is not considered as a logged user (doesn't appear as such) and is considered a "guest", like the rest of people connecting in the office. I say that because of this: http://social.technet.microsoft.com/Forums/windowsserver/en-US/ff6d3726-ff41-4d3f-9d97-5361af0206dd/vpn-users-on-server-shows-as-guest?forum=winserverNIS Also, because when I create a txt file using the VPN connection, owner field shows in description as "guest". Am I right? How can I set different rights for the VPNuser from the rest of "guest" users in the office?

    Read the article

  • Update a war on Tomcat start up

    - by pater
    I want to setup an update process for an application running on Tomcat. The server which hosts tomcat is only open during working hours (it is an intraner application for a small company). I was thinking that I could upload the new war to the server and set up "something" to run on the next server boot. This something could be a bat file that will be executed on server start up but before the start up of the Tomcat service and it will delete the old war and its exploded folder. When I update manually the war I also delete the work folder of Tomcat (just to be sure). I know about hot deployment but I do not consider it an option since I am not very sure for the implications it might have on the users current working sessions. Is there a way to run such a bat file before Tomcat start up or an alternative way to do this update? Tomcat version isn't an issue. Now is running Tomcat 6 but I can upgrade to version 7 if needed.

    Read the article

  • What happens when you add a Graphics card to a i7 with build in graphics (e.g. HD 4000)

    - by Matt
    I'm thinking of upgrading my computer from an AMD Phenom II X4 955BE with a AMD Radeon HD 6800 Graphics card (not integrated) to using a Intel Core i7 3770. As I have no knowledge of integrated graphics, my question is, what happens to the computing power when not using the HD4000 integrated graphics of the CPU (does it mean the CPU will run faster then it would if I relied on it?) Also what is better the CPUs inbuilt HD4000 or my Radeon Graphics card? I am mostly interested in terms of content creation: Using Adobe After Effects, 3D Rendering etc. Not too bothered about gaming performance. I will be using the spare parts of this build and older systems to make a second computer for network renders so what will be the advantages of keeping the Radeon with the current system for that?

    Read the article

  • Older raid controllers in raid 5 vs. Jbod and SW raid

    - by TEB
    Hi. Im in the fortunate position to have 6 Supermicro older VOD servers with the following config: Supermicro 3U case, 3xPSU Dual Xeon 3ghz P4 class cpu (5 years old.. havnt checked the exact type) 4GB Ram 3ware 9500-8 SATA controller 8 SATA SLOTS and alot of free drives. 2GB FLASH Bootdrive What im curious about is the RAID5 performance on these old beasts in HW mode vs. SW on Linux with the controller set in JBOD mode. Im thinking on using Centos 5.5 or Ubuntu or ZFS RaidZ on Opensolaris. Any tips? or reccomendations ? best regards TEB

    Read the article

  • Browser-based Operating System

    - by Ross Peoples
    I have a bunch of touchscreen machines that I want to display a webpage on and have users interact with the webpage via the touchscreen. Right now, this is done with a full-blown OS with a browser set to run at startup. I think maybe the ideal solution is to use a Linux-based OS that boots up, starts X, then starts a web browser (Chrome, Firefox, or whatever) in full screen mode. What kind of options do I have? I really want to avoid using a full-blown OS like I do now. It looks unprofessional and takes a while to boot up. I was thinking maybe Chrome OS or something, but I wouldn't know how to set it up for my purposes, since it's still designed to be used as a desktop OS instead of a kiosk-type OS.

    Read the article

  • Mirroring MySQL server with diffrent configuration

    - by HTF
    I have to migrate MySQL server to a different data centre so I would like to create another MySQL slave server in new DC and then promote it to a master later on. I previously used LVM snapshots and Percona Xtrabackup for this purpose but this time I've optimized MySQL configuration file that prevents me from using these methods. Old server (backup): innodb_log_file_size = 256M innodb_log_files_in_group = 3 New server (restore): innodb_log_file_size = 512M innodb_log_files_in_group = 2 The Xtrabackup script and LVM snapshots copy the whole directory structure so the MySQL server won't start because there is a different size for InnoDB logs. Is there any solution to avoid a downtime in this case? I can't use mysqldumps as there is around 8000 databases so I would have to take the server down for a couple of hours. I was also thinking to use the old settings with Xtrabackup and then change it once the new server is promoted to a master - less downtime but I'm not sure if this will work? Thank you Regards

    Read the article

  • Secure Apache PHP vhost configuration

    - by jsimmons
    I'm looking to secure some websites running under apache using suexec. At the moment php is executed with the user/group of the file being executed. This seems to me, not secure enough. It stops vhosts interfering with each other, but does not stop malicious code writing anywhere in the vhost being used. I was thinking that a possibility would be to run scripts as nobody/vhost group, that way the vhost user could still have full access to the vhost directories, but executing php would only be able to write to files with g+w, and to execute files with g+x. This I think should stop arbitrary writing in the web dir from compromised php. Just wondering if this is crazy, ridiculous, stupid? Of course this would be done on top of existing security measures.

    Read the article

  • Why do I get "Permission denied (publickey)" when trying to SSH from local Ubuntu to a Amazon EC2 se

    - by Vorleak Chy
    I have an instance of an application running in the cloud on Amazon EC2 instance, and I need to connect it from my local Ubuntu. It works fine on one of local ubuntu and also laptop. I got message "Permission denied (publickey)" when trying to access SSH to EC2 on another local Ubuntu. It's so strange to me. I'm thinking some sort of problems with security settings on the Amazon EC2 which has limited IPs access to one instance or certificate may need to regenerate. Does anyone know a solution?

    Read the article

  • How to address an EC2 instance from both inside and outside datacenter?

    - by Alexandr Kurilin
    I'm trying to find a good way of being able to address my EC2 database instance from both inside and outside of the datacenter. Other EC2 instances need to be able to call into it, and other clients like pgAdmin might need to connect to it from the outside world as well. It's my understanding that using the internal and external DNS names is sustainable long term as each reboot leads to a change. I'm thinking of associating an Elastic IP with the instance and giving it an A record (say db1.mydomain.com) which I then will use both within and outside the datacenter. Further instances in the same role will get the same treatment and a DNS record of db2.mydomain.com etc. Now, is there a cleaner and more stable way of achieving this result? Am going about this the wrong way? Suggestions?

    Read the article

  • help building a PC that can image a dozen hard drives simultaneously

    - by Bigbio2002
    Not sure if this belongs on here or SuperUser, but here goes... I'm trying to figure out how to make a mass hard drive imaging PC out of COTS parts. A dedicated imaging device can do 10 drives at a time, but costs several thousand dollars. So far, I'm thinking to use several 3-port PCI-E Firewire cards, and use some kind of Firewire-to-IDE adapter to connect the drives themselves. The "software" would consist of scripting diskpart, or some other imaging utility. The problem is that I can't seem to find any sort of adapter. I could use standard external hard drive bays, but then I'd have a dozen power cables that I need to plug in. Ugly, messy, and inefficient. I picked Firewire over USB not only for better transfer speeds, but also because FW can deliver power over the bus (and could theoretically power a hard drive). Does anyone have any input on this?

    Read the article

  • CPU Affinity on ARM processors

    - by dsljanus
    I am using some RaspberryPI boards for a data acquisition system. They are nice boards, with plenty of community support around them, but they are really slow. I am thinking of gradually replacing them with ODROID multicore boards, with the Samsung Exynos processors. I have some experience using taskset to set CPU affinity on my servers because I am always running Node.js applications that are by definition single threaded. Now, is it possible to do this on an ARM board? I do not see why it would not in theory, but I have doubts over how well it is going to work. Does anyone have experience with this kind of hack? Also, I would appreciate any comments about ARM CPUs and how they differ from x86.

    Read the article

  • update terminal title from standard output of long running command?

    - by Sam Hasler
    I'd like to change the title of a terminal window during a long running command (for example: git svn fetch) with values greped from the output, whilst still writing to standard output. Is this possible using named pipes or tee and xargs? I'm thinking something like git svn fetch | sed "s/^\(r\d*\).*$/ \"\\\033]0;\"\1\"\\\007\"/" | xargs -l1 echo -ne based on: http://tldp.org/HOWTO/Xterm-Title-3.html Update: getting this to work would be enough: (echo "r9" ; echo "r10") | sed "s/^\(r\d*\).*$/ \"\\\033]0;\"\1\"\\\007\"/" | xargs -l1 echo -ne Update 2: This almost does what I want. I see r10, but not r9: (echo "r9" ; sleep 1 ; echo "r10") | sed "s/^\(r[0-9]*\)\.*$/\\\033]0;\1\\\007/" | xargs -0 echo -ne

    Read the article

  • How do I leverage the power of my desktop from my laptop?

    - by TheDeeno
    I have a powerful desktop, and a decent laptop. My laptop is my primary machine. When I'm home, I'd like to "tap into" the power of my desktop while using my laptop. I'm not thinking a solution like "remoting in" and actually using my desktop from my laptop. I mean sending work to the desktop when my laptop is taxed in a seamless way. Is this possible at all? How would I do it? I'll do whatever is necessary to my desktop to enable this if it's feasible. My laptop has Vista installed.

    Read the article

  • FTP error when doing file transfer

    - by Ernie
    I'm running vsftpd version 3.0.2 over FTPeS, and I'm having a bit of trouble with file transfers. It seems to work fine when I'm on the LAN, but not from an external IP address. I have the control port and data ports open on my server's software firewall and my router's firewall. When I'm using the service from an external IP address, it seems like sometimes a file transfer will complete, but it times out and I always get the client error: "426 Failure writing network stream". I've tried several clients. I'm thinking there is some sort of data sabotage either at the router or some server policy; maybe because I'm using passive ftp? Suggestions?

    Read the article

  • Naming standard for additional A records/IP addresses for IIS servers?

    - by serialhobbyist
    When you're adding another IP address to and IIS server, what naming standards do you use for the A records? Background: I've a bunch of sites on an IIS server which use (CNAME'd) host-headers and a single IP address. Server names (and A records) adhere to unfriendly (as in difficult-to-remember) naming standards whereas CNAMEs, and therefore host-headers, can be friendly. Now I've a need for several SSL certificates for different sites. I was thinking about using an additional IP address for each to-be-SSL'd site but still using friendly CNAMEs. So then I come to what to call the A record. What do you do? Related to this question.

    Read the article

  • Nexenta under KVM?

    - by Nick
    I have an Ubuntu Server running KVM. I'd like to get the benefits of ZFS so I was thinking of installing a virtual machine under KVM running Nexenta (or NexentaStor), allowing that virtual machine to have raw access to a couple of physical hard disks, and then having it share its file system with NFS so that Ubuntu can access it. I've never tried setting up KVM so that the virtual machine has access to physical drives. Does this sound feasible, and is there anything I need to watch out for? Has someone already documented something like this? Does Nexenta/ZFS function basically as well in the virtual environment as if they were running base bones? I can take a small performance hit, but I don't want it to not be as reliable because of the virtualization. Thanks.

    Read the article

  • why do Vagrant docs suggest using public IP address 33.33.33.10 for local VMs?

    - by Gert
    I'm following a tutorial to set up vagrant (a tool to build and configure portable virtual machine images), and it seems that vagrant documentation suggests using IPv4 address 33.33.33.10 to configure a new box. That is a publicly routed IP adress, so I'm a bit confused why using this address is suggested. Since I don't own this network, I should not use an address from the 33.33.33.10/8 range. Am I correct in thinking that I should only use either a public address from a network I own, or an address from one of the private ranges as defined in RFC 1918? If so, why does Vagrant documentation suggest otherwise?

    Read the article

  • FDE with DiskCryptor V1 (TRIM Support) on Crucial M4 SSD on W7 x64 - Unpartitioned Space?

    - by JamesM
    I have a new Alienware m11 laptop with a brand new Crucial M4 128GB SSD. I have installed the SSD but not used it yet. I am thinking of using FDE with DiskCryptor 1.0.732.111. I will install Win7 Pro x64. I have read about support for TRIM in v1.0.732.111 but also about leaving unpartitioned space. My SSD drive, 'Crucual M4' does not have a manufacturer built-in reserve like other drives. My questions are: Should I leave some free unpartitioned space with v1 of DC and W7x64 even though it will support TRIM or should I not do this? Should I install DC v1 first or after installing Windows 7 (assuming it is brand new, never been used SSD)?

    Read the article

  • What you'd need to setup BBS?

    - by raspi
    What I need to setup BBS nowadays? I'm thinking of BBBS or PCBoard (no telnet! too new technology). What I've thinked so far, I'd need: virtual machine which runs DOS and hook that virtual COM-port to somekind of virtual VoIP modem software somehow (is there any?). How you can call to it across internet? Can you use HyperTerminal straight with that virtual/real modem? Or will VoIP just garble the modem data and nothing will work?

    Read the article

  • How to rename everything matching a certain string in a folder

    - by lostiniceland
    Hello Everyone I am running Linux and I have some basic console knowledge but my current problem is quite difficult and I dont know how to achieve this. I want/need to rename everything within a folder that matches a given string. By everything I mean folders/files content within a file content in hidden files Basically I want to refactor a Java-project. Sure, I could use Eclipse to handle the replacing, but this leaves out the folders or resources outside of my workspace. I was thinking of a script that could do the job for me but this seems rather tricky. For instance when it comes to folder-/file-rename I want to replace only the part of the name that matches my string, the rest should remain untouched. Maybe someone already has something like this in his/her script-collection :-) Thanks in advance Marc

    Read the article

  • let mysql use full resources

    - by shekhar
    We have a Windows server 2008 R2. It has 24GB RAM and 2.926 Mhz, 8 Core(s), 8 Logical Process. We have a large MySQL database and it's taking a lot of time to execute some queries. But while running queries, I have observed it's NOT utilizing full resources. I am thinking that if I let MySQL utilize maximum resources on my server, it can reduce execution time. How can I let my MySQL server utilize maximum resources in a healthy way?

    Read the article

  • ext4: error loading journal

    - by cloudyOutside
    I have an external hard drive with two partitions: A small FAT32 which is mostly empty and works fine and a large ext4 with tons of data, most of which isn't backed up. The ext4 is visible, but can't be mounted. I get an "error loading journal" error. The drive is a Western Digital Caviar Blue 500GB. Roughly 30GB of that is FAT32 and the rest is the ext4. The light on the enclosure turns red when reading from the bad partition. It was made by Cavalry. There wasn't any warning, but coincidentally, I've been thinking lately that I should get two large capacity drives for real backups. Is there anything that can be done? I'm not even sure I have enough storage to backup everything even if it is redeemable.

    Read the article

  • Quickly translate a word from English

    - by licorna
    I'm always reading English, but I'm a Spanish native speaker (I'm working on my thesis). Sometimes I need to translate a word into Spanish, and what I do now is to open a new tab and go to Google Translate and then put the word into the input field. Just a quick translation, one word or a small phrase. I'm a Mac and Firefox user. Is there a better way to achieve this? I was thinking that maybe a dashboard widget would do the trick and I was looking for one. The other option is to install the Google Toolbar, but I really hate toolbars. I don't know, a good Firefox extension maybe?

    Read the article

  • What RAID level for a backup server?

    - by ispirto
    I'm building a server with 12 x 3TB disks to use daily backups. I'm thinking to use RAID50 to get a good 27TB usable space. The disks will be used brutally to backup 9 servers with 1.5TB of data once a day. I'll keep the backups for 2 days. So for each server I'll have 3TB of separate partitions. Do you think this kind of huge backups would stress the disks too much and make them fail? Should I better go with RAID10? Oktay

    Read the article

< Previous Page | 125 126 127 128 129 130 131 132 133 134 135 136  | Next Page >