Search Results

Search found 16278 results on 652 pages for 'fuzzy search'.

Page 467/652 | < Previous Page | 463 464 465 466 467 468 469 470 471 472 473 474  | Next Page >

  • How do I unmount a charging device in Linux?

    - by PeanutsMonkey
    I have a mobile device attached via USB to a Linux box and wish to unmount it. I ran the command fdisk -l however it does not list a mount point. I then ran the command lsusb which yielded the screenshot below. I then proceeded to search the /dev/disk/by-id directory and was found the following file The file is a symbolic link to what appears to be /dev/sdc Questions Why does it not appear when I run the command fdisk -l? How do I unmount it properly without simply yanking the USB cord from the USB port?

    Read the article

  • How to connect with MySQL server if it won't connect via the socket?

    - by cwd
    I have an account on a shared server. I have jailshell access and also PhpMyAdmin. I want to run mysql commands via SSH but I'm getting an error: $ mysql -u mySqlUser -p mySqlPw Can't connect to local MySQL server through socket '/var/lib/mysql/mysql.sock' I can connect with PHP and phpMyAdmin, so would it be possible to call mysql from the shell and have it connect via an ip and port instead of the socket? The file /var/lib/mysql/mysql.sock does not exist - maybe that is intentional, and the only thing in /etc/my.cnf is [mysqld] skip-innodb More Info I don't have access to change system settings. I did a search in /var for mysql.sock but found nothing. However, phpMyAdmin might be connecting via a socket somehow: Really it would just be great if I could connect via IP. Also tried these two syntaxes: $ mysql -u mySqlUser -p mySqlPw -h localhost $ mysql -u mySqlUser -p mySqlPw -h localhost -P 3306 Both with the same result: ERROR 2002 (HY000): Can't connect to local MySQL server through socket '/var/lib/mysql/mysql.sock' (2)

    Read the article

  • How to make Google recognize language for a multilingual website?

    - by Julien Fouilhé
    Few weeks ago, I implemented translation functionality for the website of my company. The website is now available in french and english and I did look on the internet the best way to do if we want to do not lose any ranking and to have our pages on Google. Here is what I did: I did set a response header: Content-Language:en and Content-Language:fr My URLs are formatted as: http://www.website.com/en/... and http://www.website.com/fr/... My html tag is set with a lang attribute: <html lang="en"> and <html lang="fr"> There is a <link rel="alternate" hreflang="en" href="EnglishPageUrl"> on french pages and a <link rel="alternate" hreflang="en" href="frenchPageUrl"> on english pages. But Google keeps referring to some english pages when I'm doing a search on french engine, knowing that the website was first only available in english. Is that normal? Do I have to wait still, it has been now almost one month, I thought it would be okay...? Thank you.

    Read the article

  • Windows 8.1 & manual removal of "Sentinel Runtime Drivers"

    - by uos??
    I'm trying to install Windows 8.1 (basic upgrade from Windows 8). It is complaining of Sentinel Runtime Drivers incompatibility. I've looked for it in my Programs & Features list, but it's not there by that name nor SafeNet (as related by google search) nor do I see any obvious candidates for it having been bundled with another listed application. I tried the "Sentinel Runtime Driver Cleanup" program from the SafeNet website. It ran but was ineffective, even after a reboot. Do you recognize this "Sentinel Runtime Drivers"? I ran MalwareBytes too, just in case, but it found nothing.

    Read the article

  • Is browser and bot whitelisting a practical approach?

    - by Sn3akyP3t3
    With blacklisting it takes plenty of time to monitor events to uncover undesirable behavior and then taking corrective action. I would like to avoid that daily drudgery if possible. I'm thinking whitelisting would be the answer, but I'm unsure if that is a wise approach due to the nature of deny all, allow only a few. Eventually someone out there will be blocked unintentionally is my fear. Even so, whitelisting would also block plenty of undesired traffic to pay per use items such as the Google Custom Search API as well as preserve bandwidth and my sanity. I'm not running Apache, but the idea would be the same I'm assuming. I would essentially be depending on the User Agent identifier to determine who is allowed to visit. I've tried to take into account for accessibility because some web browsers are more geared for those with disabilities although I'm not aware of any specific ones at the moment. The need to not depend on whitelisting alone to keep the site away from harm is fully understood. Other means to protect the site still need to be in place. I intend to have a honeypot, checkbox CAPTCHA, use of OWASP ESAPI, and blacklisting previous known bad IP addresses.

    Read the article

  • Most efficient Implementation a Tree in C++

    - by Topo
    I need to write a tree where each element may have any number of child elements, and because of this each branch of the tree may have any length. The tree is only going to receive elements at first and then it is going to use exclusively for iterating though it's branches in no specific order. The tree will have several million elements and must be fast but also memory efficient. My plan makes a node class to store the elements and the pointers to its children. When the tree is fully constructed, it would be transformed it to an array or something faster and if possible, loaded to the processor's cache. Construction and the search on the tree are two different problems. Can I focus on how to solve each problem on the best way individually? The construction of has to be as fast as possible but it can use memory as it pleases. Then the transformation into a format that give us speed when iterating the tree's branches. This should preferably be an array to avoid going back and forth from RAM to cache in each element of the tree. So the real question is which is the structure to implement a tree to maximize insert speed, how can I transform it to a structure that gives me the best speed and memory?

    Read the article

  • ZSH - output whole history?

    - by GorillaSandwich
    I recently switched from bash to zsh. In bash one way (besides recursive search) that I used to find previously-run commands was history | grep whatever, where whatever is the bit of command I remember. In zsh, this isn't working. history returns only a few items, even though my .zsh_history file contains many entries, which I have configured it to do. How can I output my whole history, suitable for searching with grep? (Note: I started out using ryanb's dotfiles, so perhaps it's a problem with his default settings?)

    Read the article

  • samba - join domain - automatically set workgroup

    - by ftiaronsem
    Hell alltogether Since I have to do this often, I want to automate the joining to a windows domain as much as possible. While joining a domain one has to specify realm = in the /etc/smb.conf, along with some other settings like security=ads. Among these settingst there is workgroup = My question is: Is it possible to fill this field automatically by samba, while joining a domain? Normally I would just have said never, but as I tried leaving this field blank while joining a domain, i got: Failed to join domain: Invalid configuration ("workgroup" set to '', should be 'BLABLA') and configuration modification was not requested This has made me wondering whether an automatic modification is possible and if so how? A search on the internet and the man page brought no results. It would be really great, if someone could answer that. Thanks in advance ftiaronsem

    Read the article

  • SEO - different data with same title and keywords

    - by Junaid Saeed
    here is my scenario i have a website where i redirect my users basing upon the device they were using, lets say a user is visiting from an iPad, i take him directly to the page of iPad wallpapers, the user selects iPad version & i take the user to the gallery of wallpapers where the user can select & download any wallpaper. Every wallpaper is the required resolution, i have my reasons for doing this, now the thing is there are diff. resolution. versions of an image appearing one 5 diff. sections of my website, each having their own view page Now there is only one record in db.table for the image, and basing on the my consistent naming convention of the images, i pick the required image. this means when 5 different pages are generated in 5 categorized sections of the website, due to a shared DB record, the keywords, the titles and every single detail of the 5 pages is same besides the resolution of the image, and the section specific details that the page has and yeah the pages also have different paths like wallpapers.com\ipad-1\cars\Ferrari-dino.html wallpapers.com\ipad-2\cars\Ferrari-dino.html wallpapers.com\ipad-3\cars\Ferrari-dino.html wallpapers.com\ipad-4\cars\Ferrari-dino.html wallpapers.com\ipad-5\cars\Ferrari-dino.html now this is my scenario, How do Search Engines see it and how do they rank it? Is it a Good or Normal or Bad SEO practice? If bad how dangerous it is for my sites SEO? i need your comments on my scenario.

    Read the article

  • Is the BCM4306 wireless card ipv6 capable?

    - by horroricane
    I've been trying to connect with ipv6 enabled networks with my Broadcom Wireless card under Ubuntu 12.04. The wireless card model is BCM4306. $lspci $Network controller: Broadcom Corporation BCM4306 802.11b/g Wireless LAN Controller (rev 03) $Ethernet controller: Realtek Semiconductor Co., Ltd. RTL-8139/8139C/8139C+ (rev 10) I have been unsuccesful to connect through an ipv6 address, but I can still connect to a network when assigned an ipv4 address. While searching for an answer or a solution, I know the kernel can handle dealing with ipv6. So, what's left to question should be the hardware handling the connection. Unfortunately nothing comes up when I specifically try to search for information on BCM4306 ipv6 capabilities. I just tried using a wired connection to establish an ipv6 only connection to the network I'm on right now, but I got the same behavior of constant disconnections. Maybe it's not the hardware? I don't know.. I don't want to disable ipv6 on my machine as relevant networks I'll be connecting to will be using it exclusively, but I'm not sure what is wrong and which parts should I replace/fix to get this working.. Could someone please point me in a fruitful direction to get ipv6 working under Ubuntu 12.04?

    Read the article

  • How do I extract files from one tarball to another tarball in one step?

    - by Martin
    I have some fairly large tarball archives, from which I need to extract some files. I will later repack those files to transfer them to another server. Currently that is a two (multi) step process for me: mkdir ttmp tar -vxzf large.tgz -C ttmp/ --strip-components=<INT> <folder-to-be-extracted> or alternatively with wildcards mkdir ttmp tar -vxzf large.tgz -C ttmp/ --strip-components=<INT> \ --wildcards --no-anchored '*pattern*' Then I go ahead and recompress the created folder tar -vczf small.tgz ttmp/* rm -rf ttmp How can I combine these two commands into one? Like this tar -x large.tgz > tar -c small.tgz Just to show, what I already tried: Whenever I search the terms "extract" I will end up here or here or even here. When I use the term "split" I will end up here and that is definitely not what I intend to do. When I use "repack" I end up in strange places.

    Read the article

  • Postfix + Cpanel compatibility

    - by Mishari
    I am working on a VPS that has cPanel installed and I wanted to install Postfix on it. However I find that cPanel seems to be quite tightly integrated with Exim. I have done a google search but have only found vague statements that says "cPanel does not support postfix" but I haven't found any authoritative sources in this matter. So my question is, what happens if I disable exim entirely and install postfix instead? Will it break cPanel? Will everything run as normal? Update with more questions: Can I leave Exim for cPanel's internal use and use postfix for everything else? If so, what will happen to dovecot?

    Read the article

  • sa account gets locked out frequently

    - by bala
    I have an sql server 2008 installed. there are about 6 databases created and running on this server I often get message like "SA account is locked out" is there any specific reasons for this account being locked out? is there any other place I need to check the reasons ( I checked the eventviwer but could not get much information) Edit: i found this information in teh eventviewer backups. Is this the cause? SQL Server failed to communicate with filter daemon launch service (Windows error: The service cannot be started, either because it is disabled or because it has no enabled devices associated with it. ). Full-Text filter daemon process failed to start. Full-text search functionality will not be available.

    Read the article

  • Sample domain model for online store

    - by Carel
    We are a group of 4 software development students currently studying at the Cape Peninsula University of Technology. Currently, we are tasked with developing a web application that functions as a online store. We decided to do the back-end in Java while making use of Google Guice for persistence(which is mostly irrelevant for my question). The general idea so far to use PHP to create the website. We decided that we would like to try, after handing in the project, and register a business to actually implement the website. The problem we have been experiencing is with the domain model. These are mostly small issues, however they are starting to impact the schedule of our project. Since we are all young IT students, we have virtually no experience in the business world. As such, we spend quite a significant amount of time planning the domain model in the first place. Now, some of the issues we're picking up is say the reference between the Customer entity and the order entity. Currently, we don't have the customer id in the order entity and we have a list of order entities in the customer entity. Lately, I have wondered if the persistence mechanism will put the client id physically in the order table, even if it's not in the entity? So, I started wondering, if you load a customer object, it will search the entire order table for orders with the customer's id. Now, say you have 10 000 customers and 500 000 orders, won't this take an extremely long time? There are also some business processes that I'm not completely clear on. Finally, my question is: does anyone know of a sample domain model out there that is similar to what we're trying to achieve that will be safe to look at as a reference? I don't want to be accused of stealing anybody's intellectual property, especially since we might implement this as a business.

    Read the article

  • Virtualbox and getting the Locked VT-x Error when trying to use more than 1 processor on the session

    - by Dave
    I have installed the latest version of virtualbox onto my Hewlett Pakard (h8-1170uk) I have a intel i7 2600 cpu and 8 gb of ram I can get virtual box to create several sessions of different operating systems at the same time, but whenever i try to get 1 session to open up using more than 1 processor option select ( i wanted one of my session access to 2 processors ) i keep getting this error message VT-x features locked or unavailable in MSR. (VERR_VMX_MSR_LOCKED_OR_DISABLED). Result Code: E_FAIL (0x80004005) Component: Console Interface: IConsole {1968b7d3-e3bf-4ceb-99e0-cb7c913317bb} I have search many times and cannot find a option to correct this. I have check my BIOS and there are no options about VT-x or Virtualisation or anything. Am i doing something wrong ? Why does Virtualbox run fine when just using the 1 processor option ?

    Read the article

  • Is It key_buffer or key_buffer_size?

    - by user176890
    I search the internet regarding the correct variable in my.cnf file. Some said that key_buffer_size is depreciated, but some said that key_buffer_size is the correct variable in my.cnf. So, what is really the correct variable here? Is it key_buffer or key_buffer_size? I'm using ubuntu 12.04. And also I have the two key_buffer variable in my.cnf file. This is what I got after installing mysql. The first one is located under this: [mysqld] key_buffer = 16M The other one is located under this: [isamchk] key_buffer = 16M

    Read the article

  • XboxMP Helps You Find Multiplayer Games for the Xbox

    - by Jason Fitzpatrick
    Searching for the perfect versus or cooperative Xbox multiplayer game can be challenging–the game publisher says it plays four but what exactly does that mean? XboxMP catalogs games with detailed breakdowns of what the multiplayer experience actually entails. You can search first generation Xbox games, Xbox 360 games, and Xbox Live Arcade games with a variety of detailed criteria including the number of players it supports for local, system link, and online play (it may be a deal breaker, for example, that the only way to play the game with 4 players is to have two Xboxes or an Xbox live account). In addition to the multiplayer game data each game listing also includes addition information about the general game play including peripheral support and other features. Hit up the link below to check it out and, for more help finding multiplayer cooperative games, make sure to check out previously reviewed Co-Optimus. XBoxMP What Is the Purpose of the “Do Not Cover This Hole” Hole on Hard Drives? How To Log Into The Desktop, Add a Start Menu, and Disable Hot Corners in Windows 8 HTG Explains: Why You Shouldn’t Use a Task Killer On Android

    Read the article

  • How to have Windows 7 remember a password for a Domain

    - by Kelly Jones
    About eighteen months ago, I wrote a post covering how to clear saved passwords in Windows XP.  This week at work I was reminded how useful it is to not only deleted saved passwords, but to also setup wildcard credentials using this same interface. The scenario that I run into as consultant working at a client site, is that my laptop is not a member of the Windows Domain that my client uses to secure their network. So, when I need to access file shares, shared printers, or even the clients internal websites, I’m prompted for a name and password.  By creating a wildcard entry on my laptop (for the user account that the client issued to me), I avoid this prompt and can seamlessly access these resources.  (This also works when you’ve configured Outlook to access Exchange via RPC over HTTP.) How to create a credential wild card entry in Windows 7: Go to your Start Menu --> Type "user" into the Search box Click on the “Manage your credentials” in the column on the left Click on the “Add a Windows credential” link Enter the Domain (in my case my client’s domain), something like this: *.contoso.com Enter the username and password That’s it.  You should now be able to access resources in that Domain without being prompted for your name and password.  Please note: if you are required to change your password periodically for that domain, you’ll need to update your saved password as well.

    Read the article

  • Why grep -i is so slow? How to do it faster for ASCII?

    - by Vi.
    $ time lzop -d < tvtropes-index.lzo | egrep -B 5 '[Dd][eE][sS][cC][eE][nN][dD] ?[Ff][rR][oO][mM]' real 0m0.438s $ time lzop -d < tvtropes-index.lzo | egrep -B 5 'descend ?from' -i real 0m11.294s Both search insensitively. why -i so slow? How to make fast grep -i without entering things [iI][nN] [tT][hH][iI][sS] [wW][aA][Yy]? For example, perl -ne 'print if /descend ?from/i' works fast, but '-B 5' is not as trivial to implement as in grep (as well as other options).

    Read the article

  • Filter data in sheets from a master sheet

    - by sam
    I have a 'master sheet' with lots of furniture data in it, in column A there are the suppliers names. What I would like is to be able to have my master sheet with all the info and then sub sheets named by supplier; in these sub sheets I would like to reference the master sheet and pull out all of the items that are from that supplier. For example: I would have a sheet called 'Ikea' which would look in the master sheet and search the A column for all entries of 'Ikea'. If present, copy or reference that row 1:12 in the 'ikea' sheet. I would like to do it all dynamically using references rather than copying the data. Also, I would like it to auto update rather than having to run a macro to recalculate it each time. Can this be done with formulars rather than macros?

    Read the article

  • Sticky Load Balancing with AWS

    - by John Wheal
    I have just setup a load balancer with AWS for a few instances as search engine crawlers were bringing down the site (it has millions of pages). Parts of the site allow you to login so I selected: Enable Application Generated Cookie Stickiness and everything works fine. I now wonder how this will effect my SEO and the crawlers. As I selected sticky load balancing does this mean that a crawler will be stuck on one server and therefore defeat the point in the load balancer? Any recommendations will be appreciated.

    Read the article

  • Oracle Tutor - Is Anyone Reading Your Documentation?

    - by mary.keane
    If you are responsible for documenting your business practices, wouldn't it be nice to know if anyone is using the documentation? If the employees find it useful? You might want to consider surveying the users of the documentation on a regular basis. There are a number of free survey tools online (search for "free survey tools"), and you can have a survey ready in a matter of minutes. It's as simple as gathering a list of questions and a list of email addresses. For the questions, here are some suggestions. How often do you access the policy and procedure site? How useful is the site? How easy is it to navigate the site? How often are your questions answered on the site? What suggestions do you have to make the site more useful? You may want to consider just asking a few questions each month so that employees can complete the survey in less than 5 minutes (you'll get more responses). Make sure you have several comment boxes in the survey so that the employees can give suggestions. As the users of your documentation, the employees may have some terrific ideas that will enhance the usability of your policy and procedure site. It would be great to hear your suggestions for how to survey the users of your documentation. Mary R. Keane Senior Development Manager, Oracle BPM and Tutor

    Read the article

  • CSC folder data access AND roaming profiles issues (Vista with Server 2003, then 2008)

    - by Alex Jones
    I'm a junior sysadmin for an IT contractor that helps small, local government agencies, like little towns and the like. One of our clients, a public library with ~ 50 staff users, was recently migrated from Server 2003 Standard to Server 2008 R2 Standard in a very short timeframe; our senior employee, the only network engineer, had suddenly put in his two weeks notice, so management pushed him to do this project before quitting. A bit hasty on management's part? Perhaps. Could we do anything about that? Nope. Do I have to fix this all by myself? Pretty much. The network is set up like this: a) 50ish staff workstations, all running Vista Business SP2. All staff use MS Outlook, which uses RPC-over-HTTPS ("Outlook Anywhere") for cached Exchange access to an offsite location. b) One new (virtualized) Server 2008 R2 Standard instance, running atop a Server 2008 R2 host via Hyper-V. The VM is the domain's DC, and also the site's one and only file server. Let's call that VM "NEWBOX". c) One old physical Server 2003 Standard server, running the same roles. Let's call it "OLDBOX". It's still on the network and accessible, but it's been demoted, and its shares have been disabled. No data has been deleted. c) Gigabit Ethernet everywhere. The organization's only has one domain, and it did not change during the migration. d) Most users were set up for a combo of redirected folders + offline files, but some older employees who had been with the organization a long time are still on roaming profiles. To sum up: the servers in question handle user accounts and files, nothing else (eg, no TS, no mail, no IIS, etc.) I have two major problems I'm hoping you can help me with: 1) Even though all domain users have had their redirected folders moved to the new server, and loggin in to their workstations and testing confirms that the Documents/Music/Whatever folders point to the new paths, it appears some users (not laptops or anything either!) had been working offline from OLDBOX for a long time, and nobody realized it. Here's the ugly implication: a bunch of their data now lives only in their CSC folders, because they can't access the share on OLDBOX and sync with it finally. How do I get this data out of those CSC folders, and onto NEWBOX? 2) What's the best way to migrate roaming profile users to non-roaming ones, without losing vital data like documents, any lingering PSTs, etc? Things I've thought about trying: For problem 1: a) Reenable the documents share on OLDBOX, force an Offline Files sync for ALL domain users, then copy OLDBOX's share's data to the equivalent share on NEWBOX. Reinitialize the Offline Files cache for every user. With this: How do I safely force a domain-wide Offline Files sync? Could I lose data by reenabling the share on OLDBOX and forcing the sync? Afterwards, how can I reinitialize the Offline Files cache for every user, without doing it manually, workstation by workstation? b) Determine which users have unsynced changes to OLDBOX (again, how?), search each user's CSC folder domain-wide via workstation admin shares, and grab the unsynched data. Reinitialize the Offline Files cache for every user. With this: How can I detect which users have unsynched changes with a script? How can I search each user's CSC folder, when the ownership and permissions set for CSC folders are so restrictive? Again, afterwards, how can I reinitialize the Offline Files cache for every user, without doing it manually, workstation by workstation? c) Manually visit each workstation, copy the contents of the CSC folder, and manually copy that data onto NEWBOX. Reinitialize the Offline Files cache for every user. With this: Again, how do I 'break into' the CSC folder and get to its data? As an experiment, I took one workstation's HD offsite, imaged it for safety, and then tried the following with one of our shop PCs, after attaching the drive: grant myself full control of the folder (failed), grant myself ownership of the folder (failed), run chkdsk on the whole drive to make sure nothing's messed up (all OK), try to take full control of the entire drive (failed), try to take ownership of the entire drive (failed) MS KB articles and Googling around suggests there's a utility called CSCCMD that's meant for this exact scenario...but it looks like it's available for XP, not Vista, no? Again, afterwards, how can I reinitialize the Offline Files cache for every user, without doing it manually, workstation by workstation? For problem 2: a) Figure out which users are on roaming profiles, and where their profiles 'live' on the server. Create new folders for them in the redirected folders repository, migrate existing data, and disable the roaming. With this: Finding out who's roaming isn't hard. But what's the best way to disable the roaming itself? In AD Users and Computers, or on each user's workstation? Doing it centrally on the server seems more efficient; that said, all of the KB research I've done turns up articles on how to go from local to roaming, not the other way around, so I don't have good documentation on this. In closing: we have good backups of NEWBOX and OLDBOX, but not of the workstations themselves, so anything drastic on the client side would need imaging and testing for safety. Thanks for reading along this far! Hopefully you can help me dig us out of this mess.

    Read the article

  • mac osx software always order files alphabetically rather than by type.

    - by george
    i have noticed many mac applications sort the files alphabetically rather than by type. a good example would be coda by panic.com the files in the file menu are organized alphabetically, i requested for them to add the feature to organize files by type and they've said that its a finder thing. so i went in search to look at other applications to see if they were organizing by type i noticed dreamweaver cs4 had this same problem and now including dreamweaver cs5. there has to be something in the mac that does this and that i can modify. i played with spotlight and it now displays its files by type (thinking thats what i can do) but it didnt take affect in other applications. what library are these applications using to display a file menu for its files?

    Read the article

  • ata error: UnrecovData Handshk

    - by Aaron Digulla
    I just saw this message in my log (openSUSE 11.3, Linux 2.6.36.2): ata6.00: exception Emask 0x10 SAct 0x1 SErr 0x400100 action 0x6 frozen ata6.00: irq_stat 0x08000000, interface fatal error ata6: SError: { UnrecovData Handshk } ata6.00: failed command: WRITE FPDMA QUEUED ata6.00: cmd 61/58:00:9f:69:00/00:00:00:00:00/40 tag 0 ncq 45056 out res 40/00:04:9f:69:00/00:00:00:00:00/40 Emask 0x10 (ATA bus error) ata6.00: status: { DRDY } ata6: hard resetting link ata6: SATA link up 3.0 Gbps (SStatus 123 SControl 300) ata6.00: configured for UDMA/133 ata6: EH complete Two questions: Which disk caused the error? I search /sys and /proc but couldn't find a way to map ata6 to a device (/dev/sdaXY) When such an error happens, does Linux retry the disk operation? Or do I have a corrupt filesystem?

    Read the article

< Previous Page | 463 464 465 466 467 468 469 470 471 472 473 474  | Next Page >