After configuring the Routing and Remote Access Services to be a both VPN and NAT service, I cannot ping that server anymore from the external side. But when I configure it to be just a VPN server, or just a NAT server, I do able to ping it from the external side.
My configuration is like:
DC: 192.168.20.11
VPN and NUT: Nic 1: 192.168.20.1, NIC 2: 192.168.147.171
Client: 192.168.147.1
The mask is always: 255.255.255.0
I cannot understand why when configuring the VPN server to be a VPN and NAT, I cannot ping the server anymore.
I am looking for some well known algorithms that can be considered while handling very large amount of data.(Edit- By large amount of data I refer to records in a database excluding blobs). These algorithms if not in totality but in parts may be used in big web applications like Twitter, Last.fm , Amazon ,etc.
Specifically, I'm looking for names or links to such algorithms. My primary interest lies in developing a very deep understanding on working with large database records and writing efficient code for working with the same.
Windows XP, with Windows Classic theme
I've had a sporradic problem, where the task bar seems to fill up with lots of tasks, which seem to be names of services. When I open Task Manager, the Applications tab is filled with these names.
I can still continue to use my PC to an extent, but it's not possible to remove all these icons from the taskbar, and other Windows UI features (e.g. Start bar) tend to be 'hollow' so I only see the outline; meaning I have to restart. See screenshot:
The trigger seems to be clicking on the "Show Desktop" quick link when I have several applications open (e.g. Outlook, Excel). Has anyone else come across this? It's happened to me on my home machine before, but seems to happen more frequently (2-3 times a week) at work...
Hi i have some REST services created with WCF Rest Started Kit and i need to host these services in a WPF app. can someone point in the right direction?
Is it possible to use ASP.NET Dynamic Data with SubSonic 3 in-place of Linq to SQL classes or the Entity Framework? MetaModel.RegisterContext() throws an exception if you use the context class that SubSonic generates. I thought I remembered coming across a SubSonic/Dynamic Data example back before SubSonic 3 was released but I can't find it now. Has anyone been able to get this to work?
Enterprise User Security (EUS) allows Oracle Database to use users & roles stored in LDAP for authentication and authorization.Since the 11gR2 release, OUD natively supports EUS. EUS can be easily configured during OUD setup. ODSM (the graphical admin console) can also be used to enable EUS for a new suffix.
However, enabling EUS for a new suffix using command line interface is currently not documented, so here is the procedure:
Let's assume that EUS support was enabled during initial setup.Let's o=example be the new suffix I want to use to store Enterprise users. The following sequence of command must be applied for each new suffix:
// Create a local database holding EUS context infodsconfig create-workflow-element --set base-dn:cn=OracleContext,o=example --set
enabled:true --type db-local-backend --element-name exampleContext -n
// Add a workflow element in the call path to generate on the fly attributes required by EUSdsconfig create-workflow-element --set enabled:true --type eus-context
--element-name eusContext --set next-workflow-element:exampleContext -n
// Add the context to a workflow for routingdsconfig create-workflow --set base-dn:cn=OracleContext,o=example --set
enabled:true --set workflow-element:eusContext --workflow-name exampleContext_workflow -n
//Add the new workflow to the appropriate network groupdsconfig set-network-group-prop --group-name network-group --add workflow:exampleContext_workflow -n
// Create the local database for o=exampledsconfig create-workflow-element --set base-dn:o=example --set
enabled:true --type db-local-backend --element-name example -n
// Create a workflow element in the call path to the user data to generate on the fly attributes expected by EUS
dsconfig create-workflow-element --set enabled:true --set eus-realm:o=example
--set next-workflow-element:example --type eus --element-name eusWfe// Add the db to a workflow for routingdsconfig create-workflow --set base-dn:o=example --set
enabled:true --set workflow-element:eusWfe --workflow-name example_workflow -n
//Add the new workflow to the appropriate network groupdsconfig set-network-group-prop --group-name network-group --add workflow:example_workflow -n
// Add the appropriate acis for EUSdsconfig set-access-control-handler-prop \
--add global-aci:'(target="ldap:///o=example")(targetattr="authpassword")(version 3.0; acl "EUS reads authpassword"; allow (read,search,compare) userdn="ldap:///??sub?(&(objectclass=orclservice)(objectclass=orcldbserver))";)'
dsconfig set-access-control-handler-prop \
--add global-aci:'(target="ldap:///o=example")(targetattr="orclaccountstatusevent")(version 3.0; acl "EUS writes orclaccountstatusenabled"; allow (write) userdn="ldap:///??sub?(&(objectclass=orclservice)(objectclass=orcldbserver))";)'
Last but not least you must adapt the content of the ${OUD}/config/EUS/eusData.ldif
file with your suffix value then inport it into OUD.
I am torn between buying Kaseya or Labtech for my managed services company.
Kaseya is 1400 mo for 1000 agents.
the tech console is a beautiful web 2.0 web portal but only runs in IE
Labtech is 589 mo for 500 agents. the tech console is a feature rich application that caches and syncs back to the server. very snappy no load time between clicks.
Labtech and Kaseya promise the same features. Kaseya seems far more developed less bugs and just works, but Labtech looks better engineered is easier to get into and is rumored to be a potential Kaseya killer.
The popular PTS (Platform Technology Services) technical trainings for partners now include a workshop on Big Data.
First workshop will take place in Milan on July 10-12. (You can register by clicking the link below)
Oracle Big Data
Technical
Workshop
July
10-12,
2012: Cinisello Balsamo, Milan, Italy
For more info contact [email protected]
IT security organizations are spending too much on data protection for compliance but not enough on securing key trade secrets, the real crown jewels of corporate data.
Lets say i installed two named SQL Server 2008 instances e.g. A and B, will i have two services of each type e.g. two analysis service, two reporting service and so on, one service for A and the other for B? If yes, then it is known a service listens on a port number, how two same services going to listen on the port?
I hope this is clear enough. Thanks
In my last question i asked how to best send a string from one view controller to another, both which were on a navigation stack:
http://stackoverflow.com/questions/2898860/pass-string-from-tableviewcontroller-to-viewcontroller-in-navigation-stack
However I just realised I can either pass the path to the file in the app's document's folder as the first (the table view) has already accessed the data in the file should I pass viewcontroller the data to the pushed VC?
States are passing more and more data security laws, the US Senate and the House have bills meandering through Congress, securing personal information and encrypting that data is no longer optional.
My client wants to standardize address information for existing and future addresses collected for their customers, particularly the street suffixes. The application used to enter and collect address information has the street suffix separated from the address field, but it is a textbox instead of a drop down list therefore things are not standardized. I know there are some options out there to standarize data, but they would like a less expensive alternative. Are there any functions in SQL Server that I can use to standardized data?
I want to add a dependency to my service so that it only tries to start after most of the Windows services have started. Can anyone recommend such a service that is unlikely to have been disabled for security reasons on corporate sites. As my service needs the networking services up and running I want to make sure mine starts after these. This service may also be deployed in non-corporate environments so I’m looking for a suitable service that is unlikely to have been disabled by the end user.
I'm setting up a list of ping services for a portal.
It has a blog, forum, articles, restaurants, hotels and many other information, so it is far beyond a blog.
I have a list of standard ping services for WP blogs - but I do not know if this should be literally only for blogs.
My questions are:
Is it recommended to ping blog services from a portal, such as http://blogsearch.google.com/ping/RPC2?
Are there any penalties for sites that are not recognized as blogs?
Is there some list of ping services for regular websites and not only blogs?
Thanks!
I have a data.frame:
df<-data.frame(a=c("x","x","y","y"),b=c(1,2,3,4))
> df
a b
1 x 1
2 x 2
3 y 3
4 y 4
What's the easiest way to print out each pair of values as a list of strings like this:
"x1", "x2", "y1", "y2"
Corvallis and Eugene, OR - May 13, 2010 - CenterSpace Software, a leading provider of enterprise class numerical component libraries for the .NET platform, and Floragenex, an innovative genomic research services company, today announced that they have teamed up to build a genomics data analysis pipeline in the cloud.
I've got OSX Server running on a macbook, and I'm looking to run an ubuntu server VM on it via Parallels. I'm thinking about hosting all my apache inside linux, and possibly some other services. I'm curious what a viable config would be, having not done this before. I need to do bridged network right? How do I direct web traffic to the VM instead of OSX? Haven't got my head wrapped around how this works yet so any help would be appreciated.
Choosing a long term data storage medium isn';t as easy as you may think. You might imagine that the data could be burnt to CD, locked in a cupboard and that it would last forever however unfortunatel... [Author: Chris Holgate - Computers and Internet - April 02, 2010]
If I have systems that are based on realtime data, how can I ensure that all the information that is current is redundantly stored in a file? So that when the program starts again, it uses this information to initialize itself back to where it was when it closed.
I know of xstream and HSQLDB. but wasn't sure if this was the best option for data that needs to be a literal carbon copy.
Can we have ESB Toolkit 2.0 to route the incoming messages to WCF services hosted in IIS?
I'm thinking of using content based routing, validation and transformation capabilities of ESB for WCF services.
Thanks
A corrupt table in MS Access means lost time and data. It can lead to a loss of revenue or even employment. Learn how you might be able to recover most of the data when the worst happens.
When create an Oracle Database on the Amazon cloud you will need to store you database files somewhere on the EC2 cloud. There are basically three places where database files can be stored:
1. Local drive - This is the local drive that is part of the virtual server EC2 instance.
2. Elastic Block Storage (EBS) - Network attached storage that appears as a local drive.
3. Simple Storage Server (S3) - 'Storage for the Internet'.
S3 is not high speed and intended for store static document type files. S3 can also be used for storing static web page files. Local drives are ephemeral so not appropriate to be used as a database storage device. The leaves EBS which is the best place to store database files. EBS volumes appear as local disk drives. They are actually network-attached to an Amazon EC2 instance. In addition, EBS persists independently from the running life of a single Amazon EC2 instance.
If you use an EBS backed instance for your database data, it will remain available after reboot but not after terminate. In many cases you would not need to terminate your instance but only stop it, which is equivalent of shutdown. In order to save your database data before you terminate an instance, you can snapshot the EBS to S3.
Using EBS as a data store you can move your Oracle data files from one instance to another. This allows you to move your database from one region or or zone to another. Unfortunately, to scale out your Oracle RDS on AWS you can not have read only replicas. This is only possible with the other Oracle relational database - MySQL.
The free micro instances use EBS as its storage.
This is a very good white paper that has more details:
AWS Storage Options
This white paper also discusses: SQS, SimpleDB, and Amazon RDS in the context of storage devices. However, these are not storage devices you would use to store an Oracle database.
This slide deck discusses a lot of information that is in the white paper:
AWS Storage Options slideshow
Hi,
I have MOSS 2007 environment with multiple WFE servers. can any one know what is difference between "Office SharePoint Server Search" and "Windows SharePoint Services Search ".
Which service i have to start ? If i have to Start "Office SharePoint Server Search" then what is the meaning of giving "Windows SharePoint Services Search " in Central Administration.
---Keshaw