The ibdata1 file has been increased to ~2 TB and just leave a few GB of free space on the same drive. What is worse, I forgot to turn innodb_fiel_per_tabe on.
I read from SO that the only way to reduce ibdata1 is to back up-> delete ibdata1-> restore.
Now that the ibdata1 is too big, what's the good way to do it? and how much time (days ??) will it take?
I have two more free 2 TB hard drives available for backup.
Tag: database
Encryption – How to Encrypt and Decrypt Data from a Database Using PHP Sodium?
If you have information that is encrypted, you can not search it without first decrypting it.
If the information you are looking for something with is not confidential, e.g. For example, a user ID, message ID, or otherwise tangential to your confidential information, use this information as an index to your encrypted data. A popular choice is an automatically incrementing primary key.
If the information you use for your search is confidential, your index should be implemented using a cryptographic hash function. Keep in mind that while hash collisions are rare, the fact that unrelated plaintexts have conflicting hashes is a problem important Feature of the cryptographic hash functions. After you have found the candidate line, you must check it.
You are definitely Not be able to perform full-text searches of the encrypted data without decrypting each line. Create your indexes wisely.
Content database – can we use SQL Server 2016 for SharePoint 2013?
I have a SharePoint 2013 Foundation single server farm. My current SQL version is "SQL Server 2008 R2".
I have another server running SQL Server 2016 and I want to move all content databases to my SQL Server 2016 (using the Disconnect and Attach Content Database method).
- Is it possible to use SQL Server 2016 for SharePoint 2013?
- Can we use SQL Server 2016 by changing the database's compatibility level through the database properties?
- Is there another way to create a SQL Server 2016 instance for SharePoint 2013?
Or do I need to use one of the SharePoint 2013 supported SQL servers according to the Microsoft documentation (SQL Server 2014/2012/2008 R2)?
Thank you in advance.
performance – Formula of how much the speed of the server is reduced when images are stored in the database as blob
I'm interested in knowing how fast the servers are when you save images as blob files. Suppose any kind of image is smaller than 500 KB. was there any formula to see how many images are needed to reduce the database speed from 100% to 80% or 70%, depending on the type of system? Is it possible or does anyone have experience with calculating it according to the system? For any ideas on how to set the formula or where I could read about it, I would like to thank you.
br
Azure Pipeline Copy: One or more of SQL Server to SQL Server or a storage account and how the data is stored in the database
We have several data sources in the range of
- SQL Server sources
- CSV files
- HTTP API
We want to import this data into a SQL Server database (possibly even into a SQL Data Warehouse).
question
1) Should we first run a process with different pipelines to copy the different sources to a storage account (data collection)?
Then, as a second step, use polybase to get all the data in SQL Server.
more or less like that (but we ask the database directly)
items
The trickier part is that when we first store the files in a storage account, we want to read the files from the storage account and put them in SQL Server
question
2) If we choose the option of the first storage account
How do we then save the files (which format) in the storage account?
I have done some tests on raw text files and a build table (Polybase) and I have some problems with Hadoop to get the different types of data into the SQL server. Varchars work well.
question
If we want to store the data in the SQL database and keep it history over time.
Should we store all data?
- in native format?
- or convert all to varchar or nvarchar?
Sometimes the source changes. Fields are deleted, added and changed in length.
So what's a clever way to save the data? Polybase crashes if something is different!
Database – Use ansible initialize postgresql several times
Make sure this task works the first time:
- name: Initialize the Database
command: /usr/pgsql-9.6/bin/postgresql96-setup initdb
When it is run for the second time, the following error message appears:
fatal: [192.168.0.1]: FAILED! => {"changed": true, "cmd": ["/usr/pgsql-9.6/bin/postgresql96-setup", "initdb"], "delta": "0:00:00.017590", "end": "2019-12-11 06:08:49.999631", "msg": "non-zero return code", "rc": 1, "start": "2019-12-11 06:08:49.982041", "stderr": "", "stderr_lines": [], "stdout": "Data directory is not empty!", "stdout_lines": ["Data directory is not empty!"]}
How do I avoid performing this task if the database has already been initialized on the server?
magento 1.9 – Error Calling a member function beginTransaction () for a non-object when saving to the database
I'm at a loss, I tried a lot of things, but nothing seems to work to get rid of this mistake.
Fatal error: Call to a member function beginTransaction() on a non-object in /home1/powertg2/public_html/app/code/core/Mage/Core/Model/Abstract.php on line 313
What I'm trying to accomplish is to submit an RFQ form with simple data and store results in a database.
Turn on the / Rfq / etc / config.xml file
0.0.1
Power_Rfq
submit-rfq
rfq.xml
Power_Rfq_Model
rfq_mysql4
Power_Rfq_Model_Mysql4
submitrfq
Power_Rfq
Power_Rfq_Helper
Power / Rfq / controller / IndexController.php
// Save the customer to the database
$rfq = Mage::getModel('rfq/submitrfq');
$rfq->setData('first_name', $post('first_name'));
// Other set data
$rfq->save();
Power / Rfq / Model / Submitrfq.php
class Power_Rfq_Model_Submitrfq extends Mage_Core_Model_Abstract
{
public function _construct()
{
$this->_init('rfq/submitrfq');
}
}
Power / Rfq / Model / Mysql4 / Submitrfq.php
class Power_Rfq_Model_Mysql4_Submitrfq extends Mage_Core_Model_Mysql4_Abstract
{
public function _construct()
{
$this->_init('rfq/submitrfq', 'submitrfq_id');
}
}
Power / Rfq / Model / Mysql4 / Submitrfq / Collection.php
class Power_Rfq_Model_Mysql4_Submitrfq_Collection extends Mage_Core_Model_Mysql4_Collection_Abstract
{
public function _construct()
{
$this->_init('rfq/submitrfq');
}
}
This is for Magento 1.9. I can call other model functions like getCollections (); but this just does not work.
Any and all help appreciated. Many thanks.
How to fix a broken database (the plugin editor always displays & # 39; active & # 39; s)
I have a multisite. I tried to disable plugins with phpmyadmin wp_sitemeta -> active_sitewide_plugins. I think I gave a bit of the wrong value and it was broken. I was able to restore the correct settings. Except now, I can not use the plugin editor anymore.
At first, the plugin editor always shows (active)
if it is not active. Second, when I try to work, it always means Unable to communicate back with site to check for fatal errors, so the PHP change was reverted. You will need to upload your PHP file change by some other means, such as by using SFTP.
Apart from these two things, everything seems to work perfectly. wp_sitemeta -> active_sitewide_plugins
seems to display the correct values when I enable / disable plugins. Only the plugin editor is faulty.
How can I fix this? I can not really do a clean install, it would take too long to reinstall everything. I could make a partial correction if I could somehow leave all multisite settings the same.
Database Design – DB Structure: A table for all domain values?
A database that I recently started working on was designed with one main table containing all possible values, and foreign keys in other tables indicate the nature of each of these elements.
CREATE TABLE entity
(
indexid INTEGER PRIMARY KEY NOT NULL,
entitytype INTEGER,
entityname VARCHAR,
);
For example, this main table contains a field called & # 39; entityname & # 39; that may contain records such as:
- the UPC code of the article
- the short name of the article
- the long name of the article
- the position of the article
- the provider for this article
- the quantity in stock
- (and many others)
Each record in this table has an entity type number as a foreign key, which indicates the meaning of each record in another table, such as a
entitytype = 10 means that this entity is a UPC code type
entitytype = 12 means that this entity is a provider type, and so on.
Obviously, there are dozens of such types for Entities.
I find it a little uninteresting. There is this post here (see table with all domain values), in which the author suggests the approach, as I have expressed it in my example. This may not be the best way, and I would rather agree with his opinion on the matter.
I also think that it is difficult to make efficient inquiries, and I do not think that should be the case. I would rather have more tables and keep the whole thing more intuitive. I have a window of opportunity where I can remodel this and think about long-term and proven methods.
So, can this approach be seen as a good practice in the long run and I just have to take care of it?
amazon web services – Error – Enabling or Disabling IAM Database Authentication for the DB Cluster (Mule4 RDS Connector Aurora)
I created the DBCluster with AWS-CLI. and I wanted to create DB Instance within the created cluster with Mule 4 RDS Connector. and if I give the values
- DBclustername: cluster_name
- DbInstance class: db.t2.small
- DB Instance Identifier: testdbinstance
- Engine: Aurora
it gives the following error:
The requested DB Instance will be a member of a DB Cluster. Enable or disable IAM database authentication for the DB Cluster. (Service: AmazonRDS; Status Code: 400; Error Code: InvalidParameterCombination; Request ID: e2ccfd13-f684-400f-b83f-52943bea854b)
but as I create, I set IAM database authentication to false.